var/home/core/zuul-output/0000755000175000017500000000000015067247360014537 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067260411015473 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005063447115067260403017712 0ustar rootrootOct 01 15:59:05 crc systemd[1]: Starting Kubernetes Kubelet... Oct 01 15:59:06 crc restorecon[4667]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:59:06 crc restorecon[4667]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:59:06 crc restorecon[4667]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 01 15:59:07 crc kubenswrapper[4726]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 15:59:07 crc kubenswrapper[4726]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 01 15:59:07 crc kubenswrapper[4726]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 15:59:07 crc kubenswrapper[4726]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 15:59:07 crc kubenswrapper[4726]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 01 15:59:07 crc kubenswrapper[4726]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.547399 4726 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.556869 4726 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.556909 4726 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.556919 4726 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.556927 4726 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.556937 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.556947 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.556956 4726 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.556969 4726 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.556982 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.556992 4726 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557001 4726 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557010 4726 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557019 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557027 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557035 4726 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557043 4726 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557087 4726 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557099 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557108 4726 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557116 4726 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557124 4726 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557133 4726 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557141 4726 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557148 4726 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557155 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557163 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557171 4726 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557178 4726 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557195 4726 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557203 4726 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557211 4726 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557218 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557227 4726 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557235 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557245 4726 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557256 4726 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557265 4726 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557274 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557282 4726 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557293 4726 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557301 4726 feature_gate.go:330] unrecognized feature gate: Example Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557311 4726 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557321 4726 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557329 4726 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557336 4726 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557344 4726 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557351 4726 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557359 4726 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557366 4726 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557374 4726 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557381 4726 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557389 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557398 4726 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557408 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557418 4726 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557427 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557435 4726 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557443 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557452 4726 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557459 4726 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557470 4726 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557479 4726 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557488 4726 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557496 4726 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557505 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557514 4726 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557523 4726 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557532 4726 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557540 4726 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557550 4726 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.557558 4726 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558531 4726 flags.go:64] FLAG: --address="0.0.0.0" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558558 4726 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558576 4726 flags.go:64] FLAG: --anonymous-auth="true" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558588 4726 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558600 4726 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558609 4726 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558622 4726 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558633 4726 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558644 4726 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558730 4726 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558741 4726 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558752 4726 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558761 4726 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558771 4726 flags.go:64] FLAG: --cgroup-root="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558782 4726 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558791 4726 flags.go:64] FLAG: --client-ca-file="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558800 4726 flags.go:64] FLAG: --cloud-config="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558809 4726 flags.go:64] FLAG: --cloud-provider="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558819 4726 flags.go:64] FLAG: --cluster-dns="[]" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558830 4726 flags.go:64] FLAG: --cluster-domain="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558839 4726 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558848 4726 flags.go:64] FLAG: --config-dir="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558857 4726 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558869 4726 flags.go:64] FLAG: --container-log-max-files="5" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558893 4726 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558902 4726 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558912 4726 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558921 4726 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558931 4726 flags.go:64] FLAG: --contention-profiling="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558941 4726 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558951 4726 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558961 4726 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558970 4726 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558982 4726 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.558991 4726 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559000 4726 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559009 4726 flags.go:64] FLAG: --enable-load-reader="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559018 4726 flags.go:64] FLAG: --enable-server="true" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559026 4726 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559039 4726 flags.go:64] FLAG: --event-burst="100" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559086 4726 flags.go:64] FLAG: --event-qps="50" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559099 4726 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559112 4726 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559122 4726 flags.go:64] FLAG: --eviction-hard="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559134 4726 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559144 4726 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559153 4726 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559164 4726 flags.go:64] FLAG: --eviction-soft="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559174 4726 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559183 4726 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559193 4726 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559202 4726 flags.go:64] FLAG: --experimental-mounter-path="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559211 4726 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559220 4726 flags.go:64] FLAG: --fail-swap-on="true" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559228 4726 flags.go:64] FLAG: --feature-gates="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559240 4726 flags.go:64] FLAG: --file-check-frequency="20s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559249 4726 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559258 4726 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559268 4726 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559277 4726 flags.go:64] FLAG: --healthz-port="10248" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559286 4726 flags.go:64] FLAG: --help="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559295 4726 flags.go:64] FLAG: --hostname-override="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559305 4726 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559315 4726 flags.go:64] FLAG: --http-check-frequency="20s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559325 4726 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559335 4726 flags.go:64] FLAG: --image-credential-provider-config="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559344 4726 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559353 4726 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559362 4726 flags.go:64] FLAG: --image-service-endpoint="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559371 4726 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559380 4726 flags.go:64] FLAG: --kube-api-burst="100" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559389 4726 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559399 4726 flags.go:64] FLAG: --kube-api-qps="50" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559408 4726 flags.go:64] FLAG: --kube-reserved="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559418 4726 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559427 4726 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559437 4726 flags.go:64] FLAG: --kubelet-cgroups="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559446 4726 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559456 4726 flags.go:64] FLAG: --lock-file="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559466 4726 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559475 4726 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559484 4726 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559498 4726 flags.go:64] FLAG: --log-json-split-stream="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559515 4726 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559524 4726 flags.go:64] FLAG: --log-text-split-stream="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559533 4726 flags.go:64] FLAG: --logging-format="text" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559544 4726 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559553 4726 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559562 4726 flags.go:64] FLAG: --manifest-url="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559571 4726 flags.go:64] FLAG: --manifest-url-header="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559583 4726 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559592 4726 flags.go:64] FLAG: --max-open-files="1000000" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559604 4726 flags.go:64] FLAG: --max-pods="110" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559613 4726 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559623 4726 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559632 4726 flags.go:64] FLAG: --memory-manager-policy="None" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559642 4726 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559651 4726 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559660 4726 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559669 4726 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559695 4726 flags.go:64] FLAG: --node-status-max-images="50" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559705 4726 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559714 4726 flags.go:64] FLAG: --oom-score-adj="-999" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559724 4726 flags.go:64] FLAG: --pod-cidr="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559733 4726 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559747 4726 flags.go:64] FLAG: --pod-manifest-path="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559757 4726 flags.go:64] FLAG: --pod-max-pids="-1" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559766 4726 flags.go:64] FLAG: --pods-per-core="0" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559776 4726 flags.go:64] FLAG: --port="10250" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559785 4726 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559794 4726 flags.go:64] FLAG: --provider-id="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559803 4726 flags.go:64] FLAG: --qos-reserved="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559812 4726 flags.go:64] FLAG: --read-only-port="10255" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559821 4726 flags.go:64] FLAG: --register-node="true" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559830 4726 flags.go:64] FLAG: --register-schedulable="true" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559839 4726 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559855 4726 flags.go:64] FLAG: --registry-burst="10" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559865 4726 flags.go:64] FLAG: --registry-qps="5" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559874 4726 flags.go:64] FLAG: --reserved-cpus="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559888 4726 flags.go:64] FLAG: --reserved-memory="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559901 4726 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559911 4726 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559921 4726 flags.go:64] FLAG: --rotate-certificates="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559930 4726 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559939 4726 flags.go:64] FLAG: --runonce="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559948 4726 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559957 4726 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559969 4726 flags.go:64] FLAG: --seccomp-default="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559978 4726 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559987 4726 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.559997 4726 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560007 4726 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560017 4726 flags.go:64] FLAG: --storage-driver-password="root" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560026 4726 flags.go:64] FLAG: --storage-driver-secure="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560036 4726 flags.go:64] FLAG: --storage-driver-table="stats" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560044 4726 flags.go:64] FLAG: --storage-driver-user="root" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560090 4726 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560104 4726 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560116 4726 flags.go:64] FLAG: --system-cgroups="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560127 4726 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560144 4726 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560153 4726 flags.go:64] FLAG: --tls-cert-file="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560162 4726 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560174 4726 flags.go:64] FLAG: --tls-min-version="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560183 4726 flags.go:64] FLAG: --tls-private-key-file="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560192 4726 flags.go:64] FLAG: --topology-manager-policy="none" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560201 4726 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560210 4726 flags.go:64] FLAG: --topology-manager-scope="container" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560220 4726 flags.go:64] FLAG: --v="2" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560232 4726 flags.go:64] FLAG: --version="false" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560243 4726 flags.go:64] FLAG: --vmodule="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560254 4726 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.560264 4726 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560516 4726 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560528 4726 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560537 4726 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560548 4726 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560559 4726 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560568 4726 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560578 4726 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560586 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560597 4726 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560606 4726 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560615 4726 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560623 4726 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560633 4726 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560641 4726 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560649 4726 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560659 4726 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560668 4726 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560676 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560685 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560693 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560700 4726 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560709 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560716 4726 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560725 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560733 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560741 4726 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560749 4726 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560756 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560765 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560772 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560780 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560788 4726 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560796 4726 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560804 4726 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560811 4726 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560819 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560827 4726 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560846 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560857 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560866 4726 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560874 4726 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560882 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560890 4726 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560898 4726 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560905 4726 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560913 4726 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560921 4726 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560930 4726 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560938 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560945 4726 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560953 4726 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560961 4726 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560969 4726 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560979 4726 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560988 4726 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.560996 4726 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561004 4726 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561012 4726 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561021 4726 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561028 4726 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561036 4726 feature_gate.go:330] unrecognized feature gate: Example Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561044 4726 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561080 4726 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561089 4726 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561097 4726 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561105 4726 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561113 4726 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561120 4726 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561131 4726 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561146 4726 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.561156 4726 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.561169 4726 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.574713 4726 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.574770 4726 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.574927 4726 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.574964 4726 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.574977 4726 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.574989 4726 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575001 4726 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575011 4726 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575021 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575029 4726 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575037 4726 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575045 4726 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575084 4726 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575092 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575100 4726 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575108 4726 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575116 4726 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575124 4726 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575132 4726 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575140 4726 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575147 4726 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575156 4726 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575164 4726 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575171 4726 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575179 4726 feature_gate.go:330] unrecognized feature gate: Example Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575188 4726 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575196 4726 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575205 4726 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575214 4726 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575224 4726 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575234 4726 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575245 4726 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575256 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575268 4726 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575280 4726 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575293 4726 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575307 4726 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575321 4726 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575335 4726 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575348 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575359 4726 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575370 4726 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575380 4726 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575390 4726 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575400 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575410 4726 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575420 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575430 4726 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575441 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575451 4726 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575462 4726 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575473 4726 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575483 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575493 4726 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575508 4726 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575521 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575533 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575544 4726 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575555 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575570 4726 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575581 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575593 4726 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575606 4726 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575619 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575630 4726 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575643 4726 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575654 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575667 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575676 4726 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575685 4726 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575694 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575703 4726 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575711 4726 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.575724 4726 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575957 4726 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575969 4726 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575979 4726 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575988 4726 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.575998 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576007 4726 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576015 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576024 4726 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576033 4726 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576042 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576082 4726 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576092 4726 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576101 4726 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576112 4726 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576122 4726 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576131 4726 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576140 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576149 4726 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576175 4726 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576183 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576191 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576199 4726 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576208 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576216 4726 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576224 4726 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576232 4726 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576239 4726 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576247 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576257 4726 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576267 4726 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576275 4726 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576286 4726 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576296 4726 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576307 4726 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576316 4726 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576325 4726 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576333 4726 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576341 4726 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576350 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576358 4726 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576366 4726 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576374 4726 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576382 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576390 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576398 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576406 4726 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576414 4726 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576422 4726 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576430 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576438 4726 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576446 4726 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576454 4726 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576461 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576469 4726 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576477 4726 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576486 4726 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576494 4726 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576502 4726 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576510 4726 feature_gate.go:330] unrecognized feature gate: Example Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576518 4726 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576527 4726 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576538 4726 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576548 4726 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576557 4726 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576566 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576577 4726 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576586 4726 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576594 4726 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576604 4726 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576613 4726 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.576621 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.576632 4726 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.576874 4726 server.go:940] "Client rotation is on, will bootstrap in background" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.583882 4726 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.584029 4726 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.585842 4726 server.go:997] "Starting client certificate rotation" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.585884 4726 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.586199 4726 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-19 06:18:41.42365596 +0000 UTC Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.586348 4726 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1886h19m33.837314986s for next certificate rotation Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.614006 4726 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.617269 4726 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.637452 4726 log.go:25] "Validated CRI v1 runtime API" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.672919 4726 log.go:25] "Validated CRI v1 image API" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.675747 4726 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.683317 4726 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-01-15-54-19-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.683365 4726 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.708389 4726 manager.go:217] Machine: {Timestamp:2025-10-01 15:59:07.704651008 +0000 UTC m=+0.606203655 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:fd1dce9d-91f1-4195-90c9-2a6f634e684c BootID:013da104-12c7-4736-af77-a6eb329b09c9 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:81:26:41 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:81:26:41 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:81:8c:45 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:7a:d8:d8 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:36:45:2d Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:14:8a:a2 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:a6:1a:a5:c0:ec:b0 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:9e:df:78:63:36:d3 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.708896 4726 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.709133 4726 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.709829 4726 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.710175 4726 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.710243 4726 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.714101 4726 topology_manager.go:138] "Creating topology manager with none policy" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.714142 4726 container_manager_linux.go:303] "Creating device plugin manager" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.714846 4726 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.714881 4726 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.715166 4726 state_mem.go:36] "Initialized new in-memory state store" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.715307 4726 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.720249 4726 kubelet.go:418] "Attempting to sync node with API server" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.720324 4726 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.720377 4726 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.720407 4726 kubelet.go:324] "Adding apiserver pod source" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.720432 4726 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.726236 4726 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.728933 4726 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.729763 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:07 crc kubenswrapper[4726]: E1001 15:59:07.729911 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.729867 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:07 crc kubenswrapper[4726]: E1001 15:59:07.729994 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.730866 4726 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.732878 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.732929 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.732947 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.732963 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.732994 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.733017 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.733036 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.733098 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.733120 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.733172 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.733224 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.733246 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.733295 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.734102 4726 server.go:1280] "Started kubelet" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.734477 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.735169 4726 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.735163 4726 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.735961 4726 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 01 15:59:07 crc systemd[1]: Started Kubernetes Kubelet. Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.738683 4726 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.738732 4726 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.738884 4726 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 10:37:17.759200246 +0000 UTC Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.738949 4726 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2562h38m10.020253259s for next certificate rotation Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.739165 4726 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.739211 4726 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.739310 4726 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 01 15:59:07 crc kubenswrapper[4726]: E1001 15:59:07.739153 4726 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.740722 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:07 crc kubenswrapper[4726]: E1001 15:59:07.740916 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:59:07 crc kubenswrapper[4726]: E1001 15:59:07.740913 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="200ms" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.741292 4726 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.741317 4726 factory.go:55] Registering systemd factory Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.744928 4726 factory.go:221] Registration of the systemd container factory successfully Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.745255 4726 server.go:460] "Adding debug handlers to kubelet server" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.745864 4726 factory.go:153] Registering CRI-O factory Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.745888 4726 factory.go:221] Registration of the crio container factory successfully Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.745917 4726 factory.go:103] Registering Raw factory Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.745933 4726 manager.go:1196] Started watching for new ooms in manager Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.746852 4726 manager.go:319] Starting recovery of all containers Oct 01 15:59:07 crc kubenswrapper[4726]: E1001 15:59:07.747518 4726 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.217:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a694495d455aa default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-01 15:59:07.734013354 +0000 UTC m=+0.635565971,LastTimestamp:2025-10-01 15:59:07.734013354 +0000 UTC m=+0.635565971,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.766269 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.766817 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.766859 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.766907 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.766938 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.766966 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.766990 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767018 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767080 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767114 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767143 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767174 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767210 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767245 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767275 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767302 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767329 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767411 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767441 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767466 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767492 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767524 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767553 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767582 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767610 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767639 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767672 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767703 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767733 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767761 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767789 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767815 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767846 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767876 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767905 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767936 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.767996 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768026 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768090 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768119 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768146 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768173 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768200 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768228 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768264 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768293 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768324 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768352 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768380 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768410 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768437 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768465 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768505 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768537 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768569 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768598 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768626 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768651 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768677 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768705 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768737 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768763 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768791 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768820 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768846 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768874 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768904 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768935 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.768964 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.769081 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.769118 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.769155 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.769195 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.769223 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.769263 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.769297 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.769329 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773080 4726 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773150 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773185 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773214 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773245 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773274 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773302 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773338 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773366 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773395 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773424 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773452 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773479 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773512 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773541 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773575 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773607 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773704 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773736 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773787 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773817 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773844 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773873 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773903 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773931 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773959 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.773988 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774120 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774166 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774201 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774234 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774267 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774297 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774325 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774354 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774386 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774417 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774447 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774476 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774505 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774535 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774598 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774626 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774654 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774680 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774708 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774736 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774766 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774795 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774823 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774853 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774883 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774910 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774940 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774969 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.774995 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775024 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775087 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775118 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775150 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775177 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775202 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775229 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775259 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775287 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775315 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775342 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775372 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775402 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775429 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775461 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775490 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775518 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775561 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775589 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775614 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775646 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775675 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775702 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775729 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775762 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775790 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775817 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775848 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775875 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775900 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775928 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775967 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.775994 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776021 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776081 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776119 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776146 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776177 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776204 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776234 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776259 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776287 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776315 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776347 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776377 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776405 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776431 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776459 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776485 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776511 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776541 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776567 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776597 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776624 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776652 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776682 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776710 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776737 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776765 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776796 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776825 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776856 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776885 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776920 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776948 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.776974 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.777005 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.777035 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.777096 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.777125 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.777151 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.777182 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.777211 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.777239 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.777266 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.777291 4726 reconstruct.go:97] "Volume reconstruction finished" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.777308 4726 reconciler.go:26] "Reconciler: start to sync state" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.780833 4726 manager.go:324] Recovery completed Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.800521 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.802768 4726 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.805309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.805549 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.805571 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.806572 4726 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.806696 4726 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.806739 4726 kubelet.go:2335] "Starting kubelet main sync loop" Oct 01 15:59:07 crc kubenswrapper[4726]: E1001 15:59:07.806824 4726 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 01 15:59:07 crc kubenswrapper[4726]: W1001 15:59:07.807416 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:07 crc kubenswrapper[4726]: E1001 15:59:07.807497 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.807953 4726 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.807984 4726 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.808044 4726 state_mem.go:36] "Initialized new in-memory state store" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.824774 4726 policy_none.go:49] "None policy: Start" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.826254 4726 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.826300 4726 state_mem.go:35] "Initializing new in-memory state store" Oct 01 15:59:07 crc kubenswrapper[4726]: E1001 15:59:07.840477 4726 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.894428 4726 manager.go:334] "Starting Device Plugin manager" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.894734 4726 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.894749 4726 server.go:79] "Starting device plugin registration server" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.895297 4726 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.895316 4726 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.895783 4726 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.895871 4726 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.895883 4726 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.907447 4726 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.907534 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:07 crc kubenswrapper[4726]: E1001 15:59:07.907562 4726 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.908726 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.908764 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.908777 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.908916 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.909250 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.909315 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.909798 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.909818 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.909826 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.909956 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.910137 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.910178 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.910318 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.910362 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.910375 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.910660 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.910683 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.910693 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.910822 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.910921 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.910962 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.910986 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.911070 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.911090 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.911830 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.911865 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.911880 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.911838 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.911948 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.911958 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.912067 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.912136 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.912163 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.912755 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.912787 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.912798 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.912864 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.912880 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.912890 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.913066 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.913109 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.914313 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.914350 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.914365 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:07 crc kubenswrapper[4726]: E1001 15:59:07.941764 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="400ms" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.980822 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.980879 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.980920 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.980944 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.981004 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.981068 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.981162 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.981211 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.981256 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.981313 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.981357 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.981402 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.981433 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.981466 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.981495 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.996470 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.997568 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.997612 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.997630 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:07 crc kubenswrapper[4726]: I1001 15:59:07.997661 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 15:59:07 crc kubenswrapper[4726]: E1001 15:59:07.998101 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.082724 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.082819 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.082864 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.082895 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.082929 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.082962 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.082970 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.082993 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083026 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083093 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083126 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083130 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083182 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.082992 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083198 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083248 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083309 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083210 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083158 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083495 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083520 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083142 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083230 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083353 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083555 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083672 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083595 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083707 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.083734 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.084003 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.199479 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.208284 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.208335 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.208348 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.208384 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 15:59:08 crc kubenswrapper[4726]: E1001 15:59:08.209002 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.241668 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.251188 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.267899 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.285257 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.292427 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:59:08 crc kubenswrapper[4726]: W1001 15:59:08.301569 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-a7c5dd85828edc83de813f252f6ab3dd00899572c0a8818a1d8aec364c616fd8 WatchSource:0}: Error finding container a7c5dd85828edc83de813f252f6ab3dd00899572c0a8818a1d8aec364c616fd8: Status 404 returned error can't find the container with id a7c5dd85828edc83de813f252f6ab3dd00899572c0a8818a1d8aec364c616fd8 Oct 01 15:59:08 crc kubenswrapper[4726]: W1001 15:59:08.303423 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-f39c5586bede38ae179721258d5f43c89673ea9a7d6ac3079364d8407a2f2b83 WatchSource:0}: Error finding container f39c5586bede38ae179721258d5f43c89673ea9a7d6ac3079364d8407a2f2b83: Status 404 returned error can't find the container with id f39c5586bede38ae179721258d5f43c89673ea9a7d6ac3079364d8407a2f2b83 Oct 01 15:59:08 crc kubenswrapper[4726]: W1001 15:59:08.308555 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-273930a9e23ae2ce2c1df738cc3ff37859339903a60d2dd6849a1294a910b681 WatchSource:0}: Error finding container 273930a9e23ae2ce2c1df738cc3ff37859339903a60d2dd6849a1294a910b681: Status 404 returned error can't find the container with id 273930a9e23ae2ce2c1df738cc3ff37859339903a60d2dd6849a1294a910b681 Oct 01 15:59:08 crc kubenswrapper[4726]: W1001 15:59:08.315895 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-a1fbda246337f9393ac570e6f1f68a7e62d20bb84ac68328ce7b87207376bb49 WatchSource:0}: Error finding container a1fbda246337f9393ac570e6f1f68a7e62d20bb84ac68328ce7b87207376bb49: Status 404 returned error can't find the container with id a1fbda246337f9393ac570e6f1f68a7e62d20bb84ac68328ce7b87207376bb49 Oct 01 15:59:08 crc kubenswrapper[4726]: W1001 15:59:08.320863 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-9d486bc399d214829fb392076e0cdd6b3b52a3a01f8d66c6edb7f8621ae5cdb6 WatchSource:0}: Error finding container 9d486bc399d214829fb392076e0cdd6b3b52a3a01f8d66c6edb7f8621ae5cdb6: Status 404 returned error can't find the container with id 9d486bc399d214829fb392076e0cdd6b3b52a3a01f8d66c6edb7f8621ae5cdb6 Oct 01 15:59:08 crc kubenswrapper[4726]: E1001 15:59:08.342702 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="800ms" Oct 01 15:59:08 crc kubenswrapper[4726]: W1001 15:59:08.563208 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:08 crc kubenswrapper[4726]: E1001 15:59:08.563324 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.609781 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.611875 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.611927 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.611942 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.611979 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 15:59:08 crc kubenswrapper[4726]: E1001 15:59:08.612562 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Oct 01 15:59:08 crc kubenswrapper[4726]: W1001 15:59:08.670832 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:08 crc kubenswrapper[4726]: E1001 15:59:08.670972 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.735935 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.810842 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9d486bc399d214829fb392076e0cdd6b3b52a3a01f8d66c6edb7f8621ae5cdb6"} Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.812550 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a1fbda246337f9393ac570e6f1f68a7e62d20bb84ac68328ce7b87207376bb49"} Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.814029 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"273930a9e23ae2ce2c1df738cc3ff37859339903a60d2dd6849a1294a910b681"} Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.817087 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f39c5586bede38ae179721258d5f43c89673ea9a7d6ac3079364d8407a2f2b83"} Oct 01 15:59:08 crc kubenswrapper[4726]: I1001 15:59:08.820768 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a7c5dd85828edc83de813f252f6ab3dd00899572c0a8818a1d8aec364c616fd8"} Oct 01 15:59:08 crc kubenswrapper[4726]: W1001 15:59:08.970467 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:08 crc kubenswrapper[4726]: E1001 15:59:08.970552 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:59:09 crc kubenswrapper[4726]: W1001 15:59:09.106941 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:09 crc kubenswrapper[4726]: E1001 15:59:09.107024 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:59:09 crc kubenswrapper[4726]: E1001 15:59:09.143662 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="1.6s" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.413647 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.415086 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.415125 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.415138 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.415165 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 15:59:09 crc kubenswrapper[4726]: E1001 15:59:09.415736 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.736971 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.827932 4726 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56" exitCode=0 Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.828104 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56"} Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.828267 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.830183 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.830225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.830243 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.835292 4726 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8" exitCode=0 Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.835439 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.835419 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8"} Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.836896 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.836964 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.836983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.838357 4726 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4" exitCode=0 Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.838410 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4"} Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.838529 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.839894 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.839940 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.839965 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.842578 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187"} Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.842689 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b"} Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.842729 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5"} Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.842750 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c"} Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.842906 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.844462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.844529 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.844548 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.845368 4726 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5cbc61d45e451257595e0768e02c17ac376197b25423eaaadc0c256a395ea578" exitCode=0 Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.845498 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5cbc61d45e451257595e0768e02c17ac376197b25423eaaadc0c256a395ea578"} Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.846359 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.847444 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.847498 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.847518 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.849754 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.853086 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.853145 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:09 crc kubenswrapper[4726]: I1001 15:59:09.853172 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:09 crc kubenswrapper[4726]: E1001 15:59:09.869919 4726 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.217:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a694495d455aa default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-01 15:59:07.734013354 +0000 UTC m=+0.635565971,LastTimestamp:2025-10-01 15:59:07.734013354 +0000 UTC m=+0.635565971,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.735899 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:10 crc kubenswrapper[4726]: E1001 15:59:10.744986 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="3.2s" Oct 01 15:59:10 crc kubenswrapper[4726]: W1001 15:59:10.851361 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:10 crc kubenswrapper[4726]: E1001 15:59:10.851492 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.851752 4726 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c95e8700c1fc896258d1d15a49ff6364853ff7ef50445c0ba5778efd16c86fb8" exitCode=0 Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.851843 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c95e8700c1fc896258d1d15a49ff6364853ff7ef50445c0ba5778efd16c86fb8"} Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.851961 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.853533 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.853573 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.853586 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.854667 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.854679 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4179687cd0b6fa600a3efc7b5f5761a15fb6df23af0f1b0c94170bb106b3154f"} Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.855958 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.856015 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.856034 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.858481 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"565eeb4bdc62a6f23f5a056700c8677e7de8bac37bd7cc766d4b8637c7fe3f2e"} Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.858532 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7f94a8081041e5968c831dce9988a771386310ae721226c4f58baa6a2c1af47b"} Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.858556 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"dd9fe86b3d6a1acd9512663da493ea130d4112b79d2ea24e5d94eeb76e81d852"} Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.866750 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660"} Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.866801 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1"} Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.866826 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b"} Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.866851 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.868151 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.868215 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:10 crc kubenswrapper[4726]: I1001 15:59:10.868230 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.016389 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.017854 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.017897 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.017908 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.017936 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 15:59:11 crc kubenswrapper[4726]: E1001 15:59:11.018525 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Oct 01 15:59:11 crc kubenswrapper[4726]: W1001 15:59:11.254672 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:11 crc kubenswrapper[4726]: E1001 15:59:11.254762 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:59:11 crc kubenswrapper[4726]: W1001 15:59:11.277965 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:11 crc kubenswrapper[4726]: E1001 15:59:11.278163 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:59:11 crc kubenswrapper[4726]: W1001 15:59:11.358893 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:11 crc kubenswrapper[4726]: E1001 15:59:11.358995 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.735205 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.871915 4726 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c8ed80846072f8edfd40226b620e585fa8b8e2f131424b30be12d0e7d27f7c68" exitCode=0 Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.871992 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c8ed80846072f8edfd40226b620e585fa8b8e2f131424b30be12d0e7d27f7c68"} Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.872170 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.873129 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.873157 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.873168 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.877271 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.877786 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.878118 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8a44f3e74cf21390479bd02a47f4f741a44f5da6f42b615546284a79cac38a28"} Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.878152 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2"} Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.878357 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.878578 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.878613 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.878625 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.878715 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.878783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.878804 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.879601 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.879690 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:11 crc kubenswrapper[4726]: I1001 15:59:11.879703 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:12 crc kubenswrapper[4726]: I1001 15:59:12.252374 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:12 crc kubenswrapper[4726]: I1001 15:59:12.884346 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 01 15:59:12 crc kubenswrapper[4726]: I1001 15:59:12.887394 4726 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8a44f3e74cf21390479bd02a47f4f741a44f5da6f42b615546284a79cac38a28" exitCode=255 Oct 01 15:59:12 crc kubenswrapper[4726]: I1001 15:59:12.887435 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8a44f3e74cf21390479bd02a47f4f741a44f5da6f42b615546284a79cac38a28"} Oct 01 15:59:12 crc kubenswrapper[4726]: I1001 15:59:12.887508 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:12 crc kubenswrapper[4726]: I1001 15:59:12.889724 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:12 crc kubenswrapper[4726]: I1001 15:59:12.889762 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:12 crc kubenswrapper[4726]: I1001 15:59:12.889773 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:12 crc kubenswrapper[4726]: I1001 15:59:12.890297 4726 scope.go:117] "RemoveContainer" containerID="8a44f3e74cf21390479bd02a47f4f741a44f5da6f42b615546284a79cac38a28" Oct 01 15:59:12 crc kubenswrapper[4726]: I1001 15:59:12.892706 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"123c98db1c88c13a245bf8351f4b774360c7da9bc25a4af6555dec68896f88d7"} Oct 01 15:59:12 crc kubenswrapper[4726]: I1001 15:59:12.892740 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4e5c9d8c826f12a7c2e9807bedf5cdc779cc909e3aeae02fd84b3ef9803b19bc"} Oct 01 15:59:12 crc kubenswrapper[4726]: I1001 15:59:12.892750 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"48cf222d354b944a81a8f9368bca357c490b4cccd218a77c7a6b2edba3923d37"} Oct 01 15:59:12 crc kubenswrapper[4726]: I1001 15:59:12.892763 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"15fb7ef5d7d02b41f93de3d9cd0daade9f34b3ca164fa949d8727dfe253e7661"} Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.075225 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.075503 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.077141 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.077191 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.077207 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.165179 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.896843 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.898850 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e"} Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.898902 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.898945 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.899845 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.899893 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.899906 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.903403 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c8953b3e39038c3f4d5e83cff594bc7003f2881938115d0234c0b2c53cefb210"} Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.903576 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.904679 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.904739 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:13 crc kubenswrapper[4726]: I1001 15:59:13.904761 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.219611 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.221436 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.221494 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.221510 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.221540 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.689287 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.846399 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.854609 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.854799 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.856177 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.856246 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.856265 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.906483 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.906549 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.907817 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.907879 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.907896 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.907823 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.907951 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:14 crc kubenswrapper[4726]: I1001 15:59:14.907973 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:15 crc kubenswrapper[4726]: I1001 15:59:15.384003 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 01 15:59:15 crc kubenswrapper[4726]: I1001 15:59:15.420809 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:15 crc kubenswrapper[4726]: I1001 15:59:15.909295 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:15 crc kubenswrapper[4726]: I1001 15:59:15.909468 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:15 crc kubenswrapper[4726]: I1001 15:59:15.910400 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:15 crc kubenswrapper[4726]: I1001 15:59:15.910437 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:15 crc kubenswrapper[4726]: I1001 15:59:15.910449 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:15 crc kubenswrapper[4726]: I1001 15:59:15.910725 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:15 crc kubenswrapper[4726]: I1001 15:59:15.910754 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:15 crc kubenswrapper[4726]: I1001 15:59:15.910762 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.474555 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.474788 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.476135 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.476220 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.476241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.480010 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.912406 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.912581 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.912598 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.913993 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.914037 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.914068 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.914567 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.914623 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.914620 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.914643 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.914666 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:16 crc kubenswrapper[4726]: I1001 15:59:16.914689 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:17 crc kubenswrapper[4726]: E1001 15:59:17.908417 4726 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 01 15:59:18 crc kubenswrapper[4726]: I1001 15:59:18.739407 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:18 crc kubenswrapper[4726]: I1001 15:59:18.739659 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:18 crc kubenswrapper[4726]: I1001 15:59:18.740986 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:18 crc kubenswrapper[4726]: I1001 15:59:18.741045 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:18 crc kubenswrapper[4726]: I1001 15:59:18.741078 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:18 crc kubenswrapper[4726]: I1001 15:59:18.744316 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:18 crc kubenswrapper[4726]: I1001 15:59:18.917202 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:18 crc kubenswrapper[4726]: I1001 15:59:18.918637 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:18 crc kubenswrapper[4726]: I1001 15:59:18.918695 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:18 crc kubenswrapper[4726]: I1001 15:59:18.918711 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:19 crc kubenswrapper[4726]: I1001 15:59:19.712368 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:19 crc kubenswrapper[4726]: I1001 15:59:19.919965 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:19 crc kubenswrapper[4726]: I1001 15:59:19.921250 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:19 crc kubenswrapper[4726]: I1001 15:59:19.921319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:19 crc kubenswrapper[4726]: I1001 15:59:19.921333 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:22 crc kubenswrapper[4726]: I1001 15:59:22.713186 4726 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 15:59:22 crc kubenswrapper[4726]: I1001 15:59:22.713294 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 01 15:59:22 crc kubenswrapper[4726]: I1001 15:59:22.736432 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 01 15:59:23 crc kubenswrapper[4726]: I1001 15:59:23.166355 4726 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 01 15:59:23 crc kubenswrapper[4726]: I1001 15:59:23.166444 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 01 15:59:23 crc kubenswrapper[4726]: E1001 15:59:23.946746 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="6.4s" Oct 01 15:59:24 crc kubenswrapper[4726]: E1001 15:59:24.223295 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Oct 01 15:59:24 crc kubenswrapper[4726]: I1001 15:59:24.438151 4726 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 01 15:59:24 crc kubenswrapper[4726]: I1001 15:59:24.438231 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 01 15:59:24 crc kubenswrapper[4726]: I1001 15:59:24.450224 4726 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 01 15:59:24 crc kubenswrapper[4726]: I1001 15:59:24.450294 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 01 15:59:24 crc kubenswrapper[4726]: I1001 15:59:24.872457 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 01 15:59:24 crc kubenswrapper[4726]: I1001 15:59:24.872751 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:24 crc kubenswrapper[4726]: I1001 15:59:24.874124 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:24 crc kubenswrapper[4726]: I1001 15:59:24.874205 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:24 crc kubenswrapper[4726]: I1001 15:59:24.874225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:24 crc kubenswrapper[4726]: I1001 15:59:24.886227 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 01 15:59:24 crc kubenswrapper[4726]: I1001 15:59:24.933428 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:24 crc kubenswrapper[4726]: I1001 15:59:24.934512 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:24 crc kubenswrapper[4726]: I1001 15:59:24.934546 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:24 crc kubenswrapper[4726]: I1001 15:59:24.934554 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:25 crc kubenswrapper[4726]: I1001 15:59:25.491599 4726 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]log ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]etcd ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/generic-apiserver-start-informers ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/priority-and-fairness-filter ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/start-apiextensions-informers ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/start-apiextensions-controllers ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/crd-informer-synced ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/start-system-namespaces-controller ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 01 15:59:25 crc kubenswrapper[4726]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/bootstrap-controller ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/start-kube-aggregator-informers ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/apiservice-registration-controller ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/apiservice-discovery-controller ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]autoregister-completion ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/apiservice-openapi-controller ok Oct 01 15:59:25 crc kubenswrapper[4726]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 01 15:59:25 crc kubenswrapper[4726]: livez check failed Oct 01 15:59:25 crc kubenswrapper[4726]: I1001 15:59:25.491724 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:59:27 crc kubenswrapper[4726]: E1001 15:59:27.908613 4726 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.434675 4726 trace.go:236] Trace[1764188251]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 15:59:17.374) (total time: 12059ms): Oct 01 15:59:29 crc kubenswrapper[4726]: Trace[1764188251]: ---"Objects listed" error: 12059ms (15:59:29.434) Oct 01 15:59:29 crc kubenswrapper[4726]: Trace[1764188251]: [12.059692131s] [12.059692131s] END Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.434741 4726 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.473777 4726 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.474330 4726 trace.go:236] Trace[1923247939]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 15:59:17.611) (total time: 11862ms): Oct 01 15:59:29 crc kubenswrapper[4726]: Trace[1923247939]: ---"Objects listed" error: 11862ms (15:59:29.474) Oct 01 15:59:29 crc kubenswrapper[4726]: Trace[1923247939]: [11.862824839s] [11.862824839s] END Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.474365 4726 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.474895 4726 trace.go:236] Trace[2061255654]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 15:59:15.268) (total time: 14205ms): Oct 01 15:59:29 crc kubenswrapper[4726]: Trace[2061255654]: ---"Objects listed" error: 14205ms (15:59:29.474) Oct 01 15:59:29 crc kubenswrapper[4726]: Trace[2061255654]: [14.205848186s] [14.205848186s] END Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.474937 4726 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.475529 4726 trace.go:236] Trace[147705714]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 15:59:15.196) (total time: 14279ms): Oct 01 15:59:29 crc kubenswrapper[4726]: Trace[147705714]: ---"Objects listed" error: 14279ms (15:59:29.475) Oct 01 15:59:29 crc kubenswrapper[4726]: Trace[147705714]: [14.27924402s] [14.27924402s] END Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.475580 4726 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.506931 4726 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38096->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.507025 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38096->192.168.126.11:17697: read: connection reset by peer" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.717751 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.722079 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.734964 4726 apiserver.go:52] "Watching apiserver" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.738882 4726 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.739291 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.739796 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.739930 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.739825 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.739821 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.740139 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.740285 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.740327 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.740795 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.740898 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.741556 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.742551 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.744500 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.744573 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.745085 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.745430 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.745442 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.745455 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.745862 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.774702 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.775002 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.775327 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.775546 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.775902 4726 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.785140 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.791610 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.795153 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.795190 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.795208 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.795288 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:30.295265595 +0000 UTC m=+23.196818172 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.797246 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.807950 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.820336 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.831661 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.840117 4726 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.842102 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.853823 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.863978 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.872421 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876579 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876627 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876646 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876665 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876682 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876711 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876728 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876745 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876760 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876776 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876794 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876818 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876841 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876859 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876874 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876889 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876911 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876930 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.876946 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877069 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877086 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877106 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877125 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877140 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877155 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877170 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877184 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877210 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877227 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877243 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877257 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877273 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877288 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877308 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877323 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877340 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877356 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877373 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877391 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877407 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877424 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877440 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877664 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877680 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877696 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877711 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877727 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877746 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877761 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877776 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877797 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877773 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877829 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877849 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877866 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877882 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877898 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877914 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877933 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877949 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877966 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.877983 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878000 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878018 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878035 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878070 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878088 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878109 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878126 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878157 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878175 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878191 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878226 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878232 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878266 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878345 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878379 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878387 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878416 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878434 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878439 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878469 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878675 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878690 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878711 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878727 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878746 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878762 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878779 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878797 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878800 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878884 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878913 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878941 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878961 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.878980 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879006 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879026 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879027 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879066 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879092 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879113 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879139 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879161 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879182 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879201 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879224 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879237 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879241 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879286 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879307 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879330 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879348 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879369 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879388 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879408 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879427 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879445 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879464 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879513 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879532 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879550 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879574 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879593 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879609 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879627 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879646 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879665 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879685 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879704 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879723 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879742 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879798 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879823 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879847 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879877 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879903 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879930 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879952 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879974 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879249 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879238 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879443 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879792 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879842 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.879857 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.880006 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:59:30.379981751 +0000 UTC m=+23.281534518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.880995 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.880995 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881017 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881039 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881074 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881092 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881110 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881127 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881147 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881165 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881185 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881201 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881224 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881242 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881259 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881276 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881293 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881310 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881327 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881344 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881361 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881376 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881392 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881408 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881427 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881446 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881462 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881478 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881497 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881513 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881531 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881551 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881595 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881616 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881634 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881652 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881670 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881688 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881705 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881724 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881744 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881763 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881783 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881800 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881818 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881836 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881854 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881870 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881889 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881905 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881924 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881941 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881958 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881977 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881995 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.882012 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.882032 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.882551 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883152 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883176 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883195 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883215 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883234 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883254 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883298 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883322 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883343 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883391 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883413 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883431 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883463 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883481 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883521 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883544 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883566 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883620 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883634 4726 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883658 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883691 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883703 4726 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883714 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883731 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883741 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883751 4726 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883760 4726 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.885220 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881029 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.894952 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.890557 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.880079 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.880140 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.880170 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.880269 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.880285 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.880302 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.880443 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.880459 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.880608 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.880644 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.880835 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.880911 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881410 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881575 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.895693 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881593 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881724 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881829 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.882086 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.882288 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.882350 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883023 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883035 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883582 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.882031 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.883894 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.884120 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.884161 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.884471 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.884504 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.884666 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.884782 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.884625 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.885013 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.885013 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.884878 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.885537 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.885651 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.885859 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.886269 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.886575 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.886770 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.886788 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.886829 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.886922 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.886906 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.887189 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.887361 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.887461 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.887562 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.887673 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.887953 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.888104 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.888114 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.888148 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.888904 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.889517 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.889536 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.889717 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.889880 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.889956 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.890250 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.890370 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.890439 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.890662 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.890774 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.890917 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.890981 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.891036 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.891616 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.891191 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.892202 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.892396 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.892476 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.892835 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.893142 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.893194 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.893445 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.893767 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.893805 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.894264 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.894330 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.894563 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.896228 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.896250 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.896254 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.897340 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.896250 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.896272 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.896294 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.896305 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.896313 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.896320 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.896331 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.896341 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.896354 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.896366 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.897615 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.897789 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.897896 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.897917 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.898250 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.898422 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.881120 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.898417 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.899231 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.899278 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.900225 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.900361 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.900464 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.900753 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.900946 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.901282 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.901436 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.902953 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.903578 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.903776 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.899485 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.905413 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.905646 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.905651 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.905859 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.905937 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.905668 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.906094 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.906465 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.906486 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.906561 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.906567 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.906852 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.907195 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.907468 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.907648 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.907724 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.907760 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.907849 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:30.40782127 +0000 UTC m=+23.309373837 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.907903 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.908061 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.908209 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.908314 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.908315 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.908407 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:30.408389315 +0000 UTC m=+23.309941892 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.908637 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.908709 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.899898 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.908757 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.908798 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.900289 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.908949 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.908975 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.909371 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.908825 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.900430 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.909425 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.909455 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.901171 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.901406 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.901412 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.901443 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.909762 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.910574 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.911841 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.911926 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.912023 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.912183 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.912341 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.913185 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.913498 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.913716 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.913906 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.920794 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.920818 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.920821 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.920837 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.920835 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.920918 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:30.420899198 +0000 UTC m=+23.322451775 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.920929 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.920995 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.921488 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.922544 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.924733 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.926530 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.932387 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.932442 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.933069 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.933522 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.946097 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.950644 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.951321 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.952883 4726 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e" exitCode=255 Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.953099 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e"} Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.953240 4726 scope.go:117] "RemoveContainer" containerID="8a44f3e74cf21390479bd02a47f4f741a44f5da6f42b615546284a79cac38a28" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.958447 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.962577 4726 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.966695 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.968407 4726 scope.go:117] "RemoveContainer" containerID="1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e" Oct 01 15:59:29 crc kubenswrapper[4726]: E1001 15:59:29.968604 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.969123 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.969422 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.973099 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.982382 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984724 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984807 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984822 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984832 4726 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984844 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984853 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984863 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984873 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984882 4726 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984893 4726 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984902 4726 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984898 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984911 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984974 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984984 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.984997 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985006 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985016 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985026 4726 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985036 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985064 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985073 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985083 4726 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985092 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985102 4726 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985111 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985123 4726 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985136 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985148 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985160 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985173 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985183 4726 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985194 4726 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985204 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985214 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985223 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985232 4726 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985243 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985254 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985264 4726 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985273 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985283 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985292 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985301 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985311 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985320 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985329 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985338 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985347 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985358 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985367 4726 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985375 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985383 4726 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985395 4726 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985403 4726 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985412 4726 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985420 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985428 4726 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985438 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985449 4726 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985462 4726 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985475 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985487 4726 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985500 4726 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985515 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985527 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985540 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985557 4726 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985571 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985622 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985635 4726 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985647 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985661 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985674 4726 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985686 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985699 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985712 4726 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985724 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985735 4726 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985746 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985757 4726 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985778 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985788 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985801 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985814 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985827 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985838 4726 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985846 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985855 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985864 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985873 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985884 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985898 4726 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985912 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985925 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985936 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985949 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985961 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985973 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985984 4726 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.985996 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986006 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986019 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986031 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986041 4726 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986073 4726 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986086 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986098 4726 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986109 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986121 4726 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986132 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986143 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986155 4726 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986167 4726 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986178 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986204 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986216 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986229 4726 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986240 4726 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986252 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986264 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986275 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986286 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986298 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986310 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986323 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986333 4726 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986351 4726 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986362 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986373 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986384 4726 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986396 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986411 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986425 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986436 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986448 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986459 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986471 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986481 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986492 4726 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986503 4726 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986515 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986527 4726 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986538 4726 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986549 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986560 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986571 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986582 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986593 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986605 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986617 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986636 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986648 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986659 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986671 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986682 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986693 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986703 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986714 4726 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986729 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986741 4726 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986752 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986762 4726 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986773 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986784 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986796 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986809 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986821 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986832 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986843 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986855 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986866 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986877 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986890 4726 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986901 4726 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986913 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986926 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986936 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986948 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986958 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986969 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.986980 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.987018 4726 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.987031 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.987043 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.987070 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:29 crc kubenswrapper[4726]: I1001 15:59:29.997304 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.013801 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.032193 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.047165 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.059221 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.062179 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.070545 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:59:30 crc kubenswrapper[4726]: W1001 15:59:30.077569 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-c37cd840d71fbc38100055b1427a773e298b45315bc63a572760fba4fd6a055d WatchSource:0}: Error finding container c37cd840d71fbc38100055b1427a773e298b45315bc63a572760fba4fd6a055d: Status 404 returned error can't find the container with id c37cd840d71fbc38100055b1427a773e298b45315bc63a572760fba4fd6a055d Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.081627 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:59:30 crc kubenswrapper[4726]: W1001 15:59:30.102818 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-28d317581f8dbe97c390913f617c22b91faa6c0624f911199d9b4a1c437812aa WatchSource:0}: Error finding container 28d317581f8dbe97c390913f617c22b91faa6c0624f911199d9b4a1c437812aa: Status 404 returned error can't find the container with id 28d317581f8dbe97c390913f617c22b91faa6c0624f911199d9b4a1c437812aa Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.391488 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.391637 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.391762 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:59:31.39171791 +0000 UTC m=+24.293270487 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.391779 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.391811 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.391826 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.391903 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:31.391880175 +0000 UTC m=+24.293432982 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.425954 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.438780 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.451486 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.462399 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.474581 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.492475 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.492517 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.492535 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.492637 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.492696 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:31.492679936 +0000 UTC m=+24.394232513 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.492705 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.492813 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.492883 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.492906 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.492845 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:31.49282172 +0000 UTC m=+24.394374357 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.493110 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:31.492996245 +0000 UTC m=+24.394548832 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.493660 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a44f3e74cf21390479bd02a47f4f741a44f5da6f42b615546284a79cac38a28\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:12Z\\\",\\\"message\\\":\\\"W1001 15:59:11.896728 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 15:59:11.897653 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759334351 cert, and key in /tmp/serving-cert-2816282059/serving-signer.crt, /tmp/serving-cert-2816282059/serving-signer.key\\\\nI1001 15:59:12.206534 1 observer_polling.go:159] Starting file observer\\\\nW1001 15:59:12.212309 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 15:59:12.212532 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:59:12.215039 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2816282059/tls.crt::/tmp/serving-cert-2816282059/tls.key\\\\\\\"\\\\nF1001 15:59:12.520204 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.529067 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.553604 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.582035 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.623619 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.625434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.625513 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.625528 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.625648 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.644589 4726 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.645109 4726 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.648333 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.648390 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.648407 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.648429 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.648440 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:30Z","lastTransitionTime":"2025-10-01T15:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.676544 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.680846 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.680894 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.680904 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.680925 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.680939 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:30Z","lastTransitionTime":"2025-10-01T15:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.691506 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.695803 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.695840 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.695851 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.695866 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.695876 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:30Z","lastTransitionTime":"2025-10-01T15:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.705569 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.708814 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.708853 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.708866 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.708880 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.708890 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:30Z","lastTransitionTime":"2025-10-01T15:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.720489 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.724436 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.724507 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.724521 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.724547 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.724568 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:30Z","lastTransitionTime":"2025-10-01T15:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.744659 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.744780 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.747078 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.747129 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.747142 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.747164 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.747176 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:30Z","lastTransitionTime":"2025-10-01T15:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.807753 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.807760 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.807905 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.807988 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.849966 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.850010 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.850020 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.850037 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.850064 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:30Z","lastTransitionTime":"2025-10-01T15:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.952788 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.952837 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.952850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.952869 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.952879 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:30Z","lastTransitionTime":"2025-10-01T15:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.956481 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.959151 4726 scope.go:117] "RemoveContainer" containerID="1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e" Oct 01 15:59:30 crc kubenswrapper[4726]: E1001 15:59:30.959322 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.959444 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"28d317581f8dbe97c390913f617c22b91faa6c0624f911199d9b4a1c437812aa"} Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.960899 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9"} Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.960946 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"700460232e0dfd890f404a1480d08a7320ff9265ab816f17056cbf2d1b27c905"} Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.962871 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0"} Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.962907 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc"} Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.962920 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c37cd840d71fbc38100055b1427a773e298b45315bc63a572760fba4fd6a055d"} Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.965244 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:30 crc kubenswrapper[4726]: I1001 15:59:30.983656 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.001770 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.026151 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.039745 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.055872 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.055922 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.055933 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.055953 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.055964 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:31Z","lastTransitionTime":"2025-10-01T15:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.056772 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.073286 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.089174 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.106457 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.121109 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.133626 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.145866 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.158558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.158608 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.158618 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.158636 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.158649 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:31Z","lastTransitionTime":"2025-10-01T15:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.160373 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.171531 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.188345 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.199175 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.212946 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.231996 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-5sdsp"] Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.232387 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-cx25v"] Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.232534 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5sdsp" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.233025 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.233648 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-h8cql"] Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.234118 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.237268 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.237313 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.238162 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.238406 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.238470 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.238669 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.238682 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.238682 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.238689 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.239426 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.249668 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.262408 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.262611 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.263754 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.264230 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.264317 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:31Z","lastTransitionTime":"2025-10-01T15:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.269940 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.283780 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.295646 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.300767 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-multus-socket-dir-parent\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.300972 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67cvk\" (UniqueName: \"kubernetes.io/projected/e9257aec-e319-4261-ae4c-dec88468b680-kube-api-access-67cvk\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301088 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e777c1f8-99c8-4dd9-a920-7bcfcdd5e229-hosts-file\") pod \"node-resolver-5sdsp\" (UID: \"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\") " pod="openshift-dns/node-resolver-5sdsp" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301269 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e9257aec-e319-4261-ae4c-dec88468b680-cni-binary-copy\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301356 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-var-lib-cni-bin\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301389 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-cnibin\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301416 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-cnibin\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301444 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkm7p\" (UniqueName: \"kubernetes.io/projected/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-kube-api-access-mkm7p\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301520 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-system-cni-dir\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301604 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-cni-binary-copy\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301712 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e9257aec-e319-4261-ae4c-dec88468b680-multus-daemon-config\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301747 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-run-netns\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301774 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-var-lib-cni-multus\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301796 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-os-release\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301831 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-multus-cni-dir\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301857 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-os-release\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301886 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-system-cni-dir\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301907 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-run-multus-certs\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301931 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-run-k8s-cni-cncf-io\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301966 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-hostroot\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.301989 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-etc-kubernetes\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.302013 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.302043 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-tuning-conf-dir\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.302099 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-var-lib-kubelet\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.302120 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-multus-conf-dir\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.302144 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v77c\" (UniqueName: \"kubernetes.io/projected/e777c1f8-99c8-4dd9-a920-7bcfcdd5e229-kube-api-access-5v77c\") pod \"node-resolver-5sdsp\" (UID: \"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\") " pod="openshift-dns/node-resolver-5sdsp" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.307939 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.321697 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.333610 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.346397 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.356589 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.366520 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.366596 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.366610 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.366652 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.366663 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:31Z","lastTransitionTime":"2025-10-01T15:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.373874 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.385672 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.402884 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.402954 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-cnibin\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.402979 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-cnibin\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.402998 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkm7p\" (UniqueName: \"kubernetes.io/projected/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-kube-api-access-mkm7p\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403025 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-system-cni-dir\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403107 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-system-cni-dir\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.403124 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:59:33.403074975 +0000 UTC m=+26.304627552 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403163 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-cnibin\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403199 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-cni-binary-copy\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403167 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-cnibin\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403301 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e9257aec-e319-4261-ae4c-dec88468b680-multus-daemon-config\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403346 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-run-netns\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403391 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-var-lib-cni-multus\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403417 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-os-release\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403458 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403478 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-run-netns\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403496 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-multus-cni-dir\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403530 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-os-release\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403568 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-var-lib-cni-multus\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403529 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-os-release\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403571 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-system-cni-dir\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.403638 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403664 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-run-multus-certs\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.403669 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403672 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-os-release\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.403689 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403716 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-system-cni-dir\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403725 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-multus-cni-dir\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403722 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-run-k8s-cni-cncf-io\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403722 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-run-multus-certs\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403688 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-run-k8s-cni-cncf-io\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.403797 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:33.403770503 +0000 UTC m=+26.305323250 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403818 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403849 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-hostroot\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403866 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-etc-kubernetes\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403885 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-multus-conf-dir\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403900 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-tuning-conf-dir\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403917 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-var-lib-kubelet\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403934 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v77c\" (UniqueName: \"kubernetes.io/projected/e777c1f8-99c8-4dd9-a920-7bcfcdd5e229-kube-api-access-5v77c\") pod \"node-resolver-5sdsp\" (UID: \"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\") " pod="openshift-dns/node-resolver-5sdsp" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403954 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-multus-socket-dir-parent\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403960 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-hostroot\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403971 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67cvk\" (UniqueName: \"kubernetes.io/projected/e9257aec-e319-4261-ae4c-dec88468b680-kube-api-access-67cvk\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403994 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e9257aec-e319-4261-ae4c-dec88468b680-cni-binary-copy\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.404010 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e777c1f8-99c8-4dd9-a920-7bcfcdd5e229-hosts-file\") pod \"node-resolver-5sdsp\" (UID: \"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\") " pod="openshift-dns/node-resolver-5sdsp" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.404026 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-var-lib-cni-bin\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.404009 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-var-lib-kubelet\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.404082 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-host-var-lib-cni-bin\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.404100 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-multus-conf-dir\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.404126 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-cni-binary-copy\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.404386 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e777c1f8-99c8-4dd9-a920-7bcfcdd5e229-hosts-file\") pod \"node-resolver-5sdsp\" (UID: \"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\") " pod="openshift-dns/node-resolver-5sdsp" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.403932 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-etc-kubernetes\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.404442 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e9257aec-e319-4261-ae4c-dec88468b680-multus-socket-dir-parent\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.404531 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.404938 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-tuning-conf-dir\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.405037 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e9257aec-e319-4261-ae4c-dec88468b680-cni-binary-copy\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.405606 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e9257aec-e319-4261-ae4c-dec88468b680-multus-daemon-config\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.418540 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.428149 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkm7p\" (UniqueName: \"kubernetes.io/projected/21c32caa-f2fd-4c2a-a766-52f3fc81a0ae-kube-api-access-mkm7p\") pod \"multus-additional-cni-plugins-cx25v\" (UID: \"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\") " pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.430378 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v77c\" (UniqueName: \"kubernetes.io/projected/e777c1f8-99c8-4dd9-a920-7bcfcdd5e229-kube-api-access-5v77c\") pod \"node-resolver-5sdsp\" (UID: \"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\") " pod="openshift-dns/node-resolver-5sdsp" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.439304 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.456907 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67cvk\" (UniqueName: \"kubernetes.io/projected/e9257aec-e319-4261-ae4c-dec88468b680-kube-api-access-67cvk\") pod \"multus-h8cql\" (UID: \"e9257aec-e319-4261-ae4c-dec88468b680\") " pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.469095 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.469139 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.469151 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.469170 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.469182 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:31Z","lastTransitionTime":"2025-10-01T15:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.471456 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.486754 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.495945 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.504830 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.504897 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.504929 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.505066 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.505117 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.505166 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.505208 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.505236 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.505203 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:33.50517411 +0000 UTC m=+26.406726687 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.505358 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:33.505325314 +0000 UTC m=+26.406878071 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.505377 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:33.505367425 +0000 UTC m=+26.406920002 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.507642 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.523225 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.537782 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.546602 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5sdsp" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.552283 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.556606 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-cx25v" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.569944 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-h8cql" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.572571 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.572615 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.572627 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.572649 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.572664 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:31Z","lastTransitionTime":"2025-10-01T15:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.596016 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-x5gdr"] Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.597191 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.600066 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.600069 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.600155 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.600240 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.601131 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.615438 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: W1001 15:59:31.618109 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21c32caa_f2fd_4c2a_a766_52f3fc81a0ae.slice/crio-28c3143002077bb241ee66abbcd8edafc2f765934ee74a0beabc23c6dcbe6c95 WatchSource:0}: Error finding container 28c3143002077bb241ee66abbcd8edafc2f765934ee74a0beabc23c6dcbe6c95: Status 404 returned error can't find the container with id 28c3143002077bb241ee66abbcd8edafc2f765934ee74a0beabc23c6dcbe6c95 Oct 01 15:59:31 crc kubenswrapper[4726]: W1001 15:59:31.621174 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9257aec_e319_4261_ae4c_dec88468b680.slice/crio-b74027a42bd28b01f321a6809cd65825b4f7141875590949dc69d33e4d819943 WatchSource:0}: Error finding container b74027a42bd28b01f321a6809cd65825b4f7141875590949dc69d33e4d819943: Status 404 returned error can't find the container with id b74027a42bd28b01f321a6809cd65825b4f7141875590949dc69d33e4d819943 Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.633564 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.652864 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.673734 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.675924 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.675983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.675996 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.676020 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.676035 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:31Z","lastTransitionTime":"2025-10-01T15:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.696550 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.707499 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/00d8cf08-6eff-420a-be98-e0dd45908be6-proxy-tls\") pod \"machine-config-daemon-x5gdr\" (UID: \"00d8cf08-6eff-420a-be98-e0dd45908be6\") " pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.707561 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/00d8cf08-6eff-420a-be98-e0dd45908be6-rootfs\") pod \"machine-config-daemon-x5gdr\" (UID: \"00d8cf08-6eff-420a-be98-e0dd45908be6\") " pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.707584 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/00d8cf08-6eff-420a-be98-e0dd45908be6-mcd-auth-proxy-config\") pod \"machine-config-daemon-x5gdr\" (UID: \"00d8cf08-6eff-420a-be98-e0dd45908be6\") " pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.707634 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jllb7\" (UniqueName: \"kubernetes.io/projected/00d8cf08-6eff-420a-be98-e0dd45908be6-kube-api-access-jllb7\") pod \"machine-config-daemon-x5gdr\" (UID: \"00d8cf08-6eff-420a-be98-e0dd45908be6\") " pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.713075 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.729024 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.744474 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.756679 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.770638 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.791299 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.794618 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.794651 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.794665 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.794687 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.794697 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:31Z","lastTransitionTime":"2025-10-01T15:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.807499 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.807646 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.808168 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jllb7\" (UniqueName: \"kubernetes.io/projected/00d8cf08-6eff-420a-be98-e0dd45908be6-kube-api-access-jllb7\") pod \"machine-config-daemon-x5gdr\" (UID: \"00d8cf08-6eff-420a-be98-e0dd45908be6\") " pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.808213 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/00d8cf08-6eff-420a-be98-e0dd45908be6-proxy-tls\") pod \"machine-config-daemon-x5gdr\" (UID: \"00d8cf08-6eff-420a-be98-e0dd45908be6\") " pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.808234 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/00d8cf08-6eff-420a-be98-e0dd45908be6-rootfs\") pod \"machine-config-daemon-x5gdr\" (UID: \"00d8cf08-6eff-420a-be98-e0dd45908be6\") " pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.808268 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/00d8cf08-6eff-420a-be98-e0dd45908be6-mcd-auth-proxy-config\") pod \"machine-config-daemon-x5gdr\" (UID: \"00d8cf08-6eff-420a-be98-e0dd45908be6\") " pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.808948 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/00d8cf08-6eff-420a-be98-e0dd45908be6-mcd-auth-proxy-config\") pod \"machine-config-daemon-x5gdr\" (UID: \"00d8cf08-6eff-420a-be98-e0dd45908be6\") " pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.809679 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/00d8cf08-6eff-420a-be98-e0dd45908be6-rootfs\") pod \"machine-config-daemon-x5gdr\" (UID: \"00d8cf08-6eff-420a-be98-e0dd45908be6\") " pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.812118 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.812811 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/00d8cf08-6eff-420a-be98-e0dd45908be6-proxy-tls\") pod \"machine-config-daemon-x5gdr\" (UID: \"00d8cf08-6eff-420a-be98-e0dd45908be6\") " pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.818500 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.819440 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.820767 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.821500 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.822647 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.823423 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.824502 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.825179 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.825417 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jllb7\" (UniqueName: \"kubernetes.io/projected/00d8cf08-6eff-420a-be98-e0dd45908be6-kube-api-access-jllb7\") pod \"machine-config-daemon-x5gdr\" (UID: \"00d8cf08-6eff-420a-be98-e0dd45908be6\") " pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.826196 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.826733 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.827255 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.828399 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.828913 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.829928 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.830509 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.831479 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.832206 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.832585 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.833588 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.834422 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.838469 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.839534 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.840042 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.842310 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.843069 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.845523 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.846433 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.847824 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.848443 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.849777 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.850458 4726 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.850607 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.853270 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.855022 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.855684 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.858444 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.859626 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.860251 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.861399 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.862071 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.862888 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.863496 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.864546 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.865179 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.866118 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.866649 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.867624 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.868392 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.869400 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.869898 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.870743 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.871288 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.871912 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.872749 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.896376 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.896413 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.896422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.896436 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.896446 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:31Z","lastTransitionTime":"2025-10-01T15:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.908893 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 15:59:31 crc kubenswrapper[4726]: W1001 15:59:31.921176 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00d8cf08_6eff_420a_be98_e0dd45908be6.slice/crio-1dcd66662dd709a452171b7052e7a6a42caf7082b2e8494cbbd745f69aa3c3f2 WatchSource:0}: Error finding container 1dcd66662dd709a452171b7052e7a6a42caf7082b2e8494cbbd745f69aa3c3f2: Status 404 returned error can't find the container with id 1dcd66662dd709a452171b7052e7a6a42caf7082b2e8494cbbd745f69aa3c3f2 Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.966539 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"1dcd66662dd709a452171b7052e7a6a42caf7082b2e8494cbbd745f69aa3c3f2"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.968091 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h8cql" event={"ID":"e9257aec-e319-4261-ae4c-dec88468b680","Type":"ContainerStarted","Data":"e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.968120 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h8cql" event={"ID":"e9257aec-e319-4261-ae4c-dec88468b680","Type":"ContainerStarted","Data":"b74027a42bd28b01f321a6809cd65825b4f7141875590949dc69d33e4d819943"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.969987 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" event={"ID":"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae","Type":"ContainerStarted","Data":"f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.970042 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" event={"ID":"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae","Type":"ContainerStarted","Data":"28c3143002077bb241ee66abbcd8edafc2f765934ee74a0beabc23c6dcbe6c95"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.972684 4726 scope.go:117] "RemoveContainer" containerID="1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e" Oct 01 15:59:31 crc kubenswrapper[4726]: E1001 15:59:31.972845 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.973029 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5sdsp" event={"ID":"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229","Type":"ContainerStarted","Data":"f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.973087 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5sdsp" event={"ID":"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229","Type":"ContainerStarted","Data":"df186535abf0b07d0af109eed2e8eecc6237c649db52e21a6e74254be1b69cbe"} Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.985383 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.992256 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zp4jh"] Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.994247 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.996775 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.996838 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.996843 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.997513 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.998260 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.998436 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.998900 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.998933 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.998945 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.998963 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:31 crc kubenswrapper[4726]: I1001 15:59:31.998975 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:31Z","lastTransitionTime":"2025-10-01T15:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.001284 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.005181 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.019330 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.035763 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.063221 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.100802 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.104490 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.104547 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.104557 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.104572 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.104584 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:32Z","lastTransitionTime":"2025-10-01T15:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111229 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovn-node-metrics-cert\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111296 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-log-socket\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111315 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znnhm\" (UniqueName: \"kubernetes.io/projected/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-kube-api-access-znnhm\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111339 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-node-log\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111362 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-run-ovn-kubernetes\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111378 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-cni-bin\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111394 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-cni-netd\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111421 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-kubelet\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111437 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-var-lib-openvswitch\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111453 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovnkube-config\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111485 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-systemd-units\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111500 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-slash\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111517 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111537 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovnkube-script-lib\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111553 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-run-netns\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111590 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-systemd\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111614 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-openvswitch\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111632 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-ovn\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111649 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-etc-openvswitch\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.111665 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-env-overrides\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.116690 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.133855 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.146775 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.167505 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.188998 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.207503 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.207554 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.207566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.207584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.207598 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:32Z","lastTransitionTime":"2025-10-01T15:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.208730 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213024 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-systemd-units\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213196 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-slash\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213129 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-systemd-units\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213308 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213380 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213375 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-slash\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213418 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovnkube-script-lib\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213454 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-run-netns\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213572 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-run-netns\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213610 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-systemd\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213641 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-openvswitch\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213764 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-ovn\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213846 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-ovn\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213689 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-systemd\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213718 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-openvswitch\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213922 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-etc-openvswitch\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213949 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-env-overrides\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.213996 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-etc-openvswitch\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.214375 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovnkube-script-lib\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.214626 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-env-overrides\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.214712 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovn-node-metrics-cert\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216307 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-log-socket\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216350 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-log-socket\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216395 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znnhm\" (UniqueName: \"kubernetes.io/projected/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-kube-api-access-znnhm\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216425 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-node-log\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216460 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-run-ovn-kubernetes\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216484 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-cni-bin\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216509 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-cni-netd\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216537 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-kubelet\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216564 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-var-lib-openvswitch\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216575 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-node-log\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216602 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-run-ovn-kubernetes\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216588 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovnkube-config\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216666 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-cni-netd\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216700 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-kubelet\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216700 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-var-lib-openvswitch\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.216803 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-cni-bin\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.217208 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovnkube-config\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.223326 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovn-node-metrics-cert\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.226137 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.245568 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.245847 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znnhm\" (UniqueName: \"kubernetes.io/projected/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-kube-api-access-znnhm\") pod \"ovnkube-node-zp4jh\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.269750 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.290724 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.306268 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.306876 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.310982 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.311020 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.311031 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.311085 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.311099 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:32Z","lastTransitionTime":"2025-10-01T15:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.323076 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.335238 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.354650 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.370571 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: W1001 15:59:32.377757 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48c2aeef_b0f7_41b1_8a03_519171bc0a6a.slice/crio-9053933dd57bfcbf4a57a5ab06bf1cb49e721929ae909665f0042374df9d604f WatchSource:0}: Error finding container 9053933dd57bfcbf4a57a5ab06bf1cb49e721929ae909665f0042374df9d604f: Status 404 returned error can't find the container with id 9053933dd57bfcbf4a57a5ab06bf1cb49e721929ae909665f0042374df9d604f Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.396011 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.414114 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.414181 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.414192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.414236 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.414256 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:32Z","lastTransitionTime":"2025-10-01T15:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.433779 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.472886 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.514507 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.516821 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.516888 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.516900 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.516936 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.516954 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:32Z","lastTransitionTime":"2025-10-01T15:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.620200 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.620248 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.620264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.620282 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.620294 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:32Z","lastTransitionTime":"2025-10-01T15:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.722836 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.722885 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.722902 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.722922 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.722933 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:32Z","lastTransitionTime":"2025-10-01T15:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.807206 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.807297 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:32 crc kubenswrapper[4726]: E1001 15:59:32.807348 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:32 crc kubenswrapper[4726]: E1001 15:59:32.807483 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.825183 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.825230 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.825243 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.825266 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.825285 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:32Z","lastTransitionTime":"2025-10-01T15:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.928440 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.928484 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.928495 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.928511 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.928521 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:32Z","lastTransitionTime":"2025-10-01T15:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.976971 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.977018 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.978571 4726 generic.go:334] "Generic (PLEG): container finished" podID="21c32caa-f2fd-4c2a-a766-52f3fc81a0ae" containerID="f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2" exitCode=0 Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.978644 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" event={"ID":"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae","Type":"ContainerDied","Data":"f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.980573 4726 generic.go:334] "Generic (PLEG): container finished" podID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerID="b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a" exitCode=0 Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.980651 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerDied","Data":"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.980710 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerStarted","Data":"9053933dd57bfcbf4a57a5ab06bf1cb49e721929ae909665f0042374df9d604f"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.982789 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386"} Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.983305 4726 scope.go:117] "RemoveContainer" containerID="1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e" Oct 01 15:59:32 crc kubenswrapper[4726]: E1001 15:59:32.983464 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 01 15:59:32 crc kubenswrapper[4726]: I1001 15:59:32.994039 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:32Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.019001 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.031887 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.031932 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.031941 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.031958 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.031969 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:33Z","lastTransitionTime":"2025-10-01T15:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.043518 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.067821 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.082935 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.100717 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.115418 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.131471 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.139695 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.139736 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.139747 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.139763 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.139775 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:33Z","lastTransitionTime":"2025-10-01T15:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.150362 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.166186 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.173123 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.187007 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.205607 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.218641 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.236706 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.242617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.242653 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.242664 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.242682 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.242693 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:33Z","lastTransitionTime":"2025-10-01T15:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.252401 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.289319 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.310713 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.323731 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.335681 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.345833 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.345873 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.345885 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.345901 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.345911 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:33Z","lastTransitionTime":"2025-10-01T15:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.351890 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.364228 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.394945 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.428612 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.428759 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.428824 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:59:37.428806473 +0000 UTC m=+30.330359050 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.428921 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.428934 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.428945 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.428990 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:37.428981368 +0000 UTC m=+30.330533945 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.431984 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.447899 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.447942 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.447953 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.447969 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.447981 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:33Z","lastTransitionTime":"2025-10-01T15:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.472721 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.516550 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.529636 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.529683 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.529704 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.529775 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.529817 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:37.52980415 +0000 UTC m=+30.431356727 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.529889 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.529900 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.529910 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.529931 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:37.529925514 +0000 UTC m=+30.431478091 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.529979 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.530004 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:37.529998185 +0000 UTC m=+30.431550762 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.548909 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:33Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.550288 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.550311 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.550322 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.550337 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.550347 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:33Z","lastTransitionTime":"2025-10-01T15:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.652874 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.652904 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.652913 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.652926 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.652934 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:33Z","lastTransitionTime":"2025-10-01T15:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.755481 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.755835 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.755846 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.755861 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.755872 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:33Z","lastTransitionTime":"2025-10-01T15:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.807484 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.807658 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.857894 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.858271 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.858363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.858477 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.858567 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:33Z","lastTransitionTime":"2025-10-01T15:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.961080 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.961120 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.961130 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.961148 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.961160 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:33Z","lastTransitionTime":"2025-10-01T15:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.988771 4726 generic.go:334] "Generic (PLEG): container finished" podID="21c32caa-f2fd-4c2a-a766-52f3fc81a0ae" containerID="6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d" exitCode=0 Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.988814 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" event={"ID":"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae","Type":"ContainerDied","Data":"6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.994196 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerStarted","Data":"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.994321 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerStarted","Data":"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.994403 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerStarted","Data":"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.994511 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerStarted","Data":"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a"} Oct 01 15:59:33 crc kubenswrapper[4726]: I1001 15:59:33.995132 4726 scope.go:117] "RemoveContainer" containerID="1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e" Oct 01 15:59:33 crc kubenswrapper[4726]: E1001 15:59:33.995264 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.003992 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.024590 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.040782 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.059135 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.066572 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.066849 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.066937 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.067094 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.067190 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:34Z","lastTransitionTime":"2025-10-01T15:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.073148 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.088393 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.099726 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.118384 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.133368 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.149520 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.167871 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.169957 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.169982 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.169990 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.170004 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.170013 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:34Z","lastTransitionTime":"2025-10-01T15:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.180207 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.208353 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.272313 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.272349 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.272359 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.272373 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.272384 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:34Z","lastTransitionTime":"2025-10-01T15:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.375832 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.375891 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.375904 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.375925 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.375940 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:34Z","lastTransitionTime":"2025-10-01T15:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.478143 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.478196 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.478211 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.478232 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.478246 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:34Z","lastTransitionTime":"2025-10-01T15:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.581390 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.581443 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.581457 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.581477 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.581493 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:34Z","lastTransitionTime":"2025-10-01T15:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.679040 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-fmbgk"] Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.679615 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-fmbgk" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.681920 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.682143 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.682782 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.683788 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.683829 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.683844 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.683864 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.683877 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:34Z","lastTransitionTime":"2025-10-01T15:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.685335 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.699135 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.711942 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.727108 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.743228 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lgm6\" (UniqueName: \"kubernetes.io/projected/bf1db398-f129-4277-8527-e9117daec3a4-kube-api-access-8lgm6\") pod \"node-ca-fmbgk\" (UID: \"bf1db398-f129-4277-8527-e9117daec3a4\") " pod="openshift-image-registry/node-ca-fmbgk" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.743299 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bf1db398-f129-4277-8527-e9117daec3a4-serviceca\") pod \"node-ca-fmbgk\" (UID: \"bf1db398-f129-4277-8527-e9117daec3a4\") " pod="openshift-image-registry/node-ca-fmbgk" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.743358 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bf1db398-f129-4277-8527-e9117daec3a4-host\") pod \"node-ca-fmbgk\" (UID: \"bf1db398-f129-4277-8527-e9117daec3a4\") " pod="openshift-image-registry/node-ca-fmbgk" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.745566 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.757867 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.770511 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.787213 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.787264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.787275 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.787293 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.787307 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:34Z","lastTransitionTime":"2025-10-01T15:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.788224 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.802407 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.807467 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.807491 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:34 crc kubenswrapper[4726]: E1001 15:59:34.807646 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:34 crc kubenswrapper[4726]: E1001 15:59:34.807854 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.816813 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.828829 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.845002 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bf1db398-f129-4277-8527-e9117daec3a4-serviceca\") pod \"node-ca-fmbgk\" (UID: \"bf1db398-f129-4277-8527-e9117daec3a4\") " pod="openshift-image-registry/node-ca-fmbgk" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.845096 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bf1db398-f129-4277-8527-e9117daec3a4-host\") pod \"node-ca-fmbgk\" (UID: \"bf1db398-f129-4277-8527-e9117daec3a4\") " pod="openshift-image-registry/node-ca-fmbgk" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.845126 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lgm6\" (UniqueName: \"kubernetes.io/projected/bf1db398-f129-4277-8527-e9117daec3a4-kube-api-access-8lgm6\") pod \"node-ca-fmbgk\" (UID: \"bf1db398-f129-4277-8527-e9117daec3a4\") " pod="openshift-image-registry/node-ca-fmbgk" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.845243 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bf1db398-f129-4277-8527-e9117daec3a4-host\") pod \"node-ca-fmbgk\" (UID: \"bf1db398-f129-4277-8527-e9117daec3a4\") " pod="openshift-image-registry/node-ca-fmbgk" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.846569 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bf1db398-f129-4277-8527-e9117daec3a4-serviceca\") pod \"node-ca-fmbgk\" (UID: \"bf1db398-f129-4277-8527-e9117daec3a4\") " pod="openshift-image-registry/node-ca-fmbgk" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.854166 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.869439 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lgm6\" (UniqueName: \"kubernetes.io/projected/bf1db398-f129-4277-8527-e9117daec3a4-kube-api-access-8lgm6\") pod \"node-ca-fmbgk\" (UID: \"bf1db398-f129-4277-8527-e9117daec3a4\") " pod="openshift-image-registry/node-ca-fmbgk" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.870290 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.881555 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.889661 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.889720 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.889735 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.889758 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.889772 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:34Z","lastTransitionTime":"2025-10-01T15:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.900161 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:34Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.993032 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.993102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.993112 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.993126 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.993136 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:34Z","lastTransitionTime":"2025-10-01T15:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:34 crc kubenswrapper[4726]: I1001 15:59:34.994141 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-fmbgk" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.004829 4726 generic.go:334] "Generic (PLEG): container finished" podID="21c32caa-f2fd-4c2a-a766-52f3fc81a0ae" containerID="2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd" exitCode=0 Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.004925 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" event={"ID":"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae","Type":"ContainerDied","Data":"2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd"} Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.011759 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerStarted","Data":"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286"} Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.011823 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerStarted","Data":"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413"} Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.021825 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:35Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.041971 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:35Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.058276 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:35Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.073932 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:35Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.085461 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:35Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.103635 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.103685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.103698 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.103717 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.103729 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:35Z","lastTransitionTime":"2025-10-01T15:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.104179 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:35Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.120066 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:35Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.134775 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:35Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.161747 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:35Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.177311 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:35Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.190313 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:35Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.206973 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.207026 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.207040 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.207081 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.207095 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:35Z","lastTransitionTime":"2025-10-01T15:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.210499 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:35Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.250475 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:35Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.295545 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:35Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.308843 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.308879 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.308890 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.308906 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.308915 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:35Z","lastTransitionTime":"2025-10-01T15:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.413169 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.413220 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.413232 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.413248 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.413261 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:35Z","lastTransitionTime":"2025-10-01T15:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.516024 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.516090 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.516102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.516121 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.516134 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:35Z","lastTransitionTime":"2025-10-01T15:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.618820 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.618882 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.618895 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.618917 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.618935 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:35Z","lastTransitionTime":"2025-10-01T15:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.721367 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.721408 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.721418 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.721434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.721447 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:35Z","lastTransitionTime":"2025-10-01T15:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.807168 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:35 crc kubenswrapper[4726]: E1001 15:59:35.807368 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.823800 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.823851 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.823865 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.823882 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.823895 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:35Z","lastTransitionTime":"2025-10-01T15:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.927435 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.927507 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.927527 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.927557 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:35 crc kubenswrapper[4726]: I1001 15:59:35.927574 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:35Z","lastTransitionTime":"2025-10-01T15:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.019012 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-fmbgk" event={"ID":"bf1db398-f129-4277-8527-e9117daec3a4","Type":"ContainerStarted","Data":"11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa"} Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.019137 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-fmbgk" event={"ID":"bf1db398-f129-4277-8527-e9117daec3a4","Type":"ContainerStarted","Data":"b0690c36db2029db7cdeb9f2e713cf311d23933b0b7115cf281b997efba10a00"} Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.023459 4726 generic.go:334] "Generic (PLEG): container finished" podID="21c32caa-f2fd-4c2a-a766-52f3fc81a0ae" containerID="384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672" exitCode=0 Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.023521 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" event={"ID":"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae","Type":"ContainerDied","Data":"384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672"} Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.030212 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.030274 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.030290 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.030886 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.030931 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:36Z","lastTransitionTime":"2025-10-01T15:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.037964 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.084938 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.126484 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.133987 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.134023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.134033 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.134066 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.134078 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:36Z","lastTransitionTime":"2025-10-01T15:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.141491 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.157891 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.169157 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.180610 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.190370 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.203068 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.214590 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.236680 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.236711 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.236723 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.236740 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.236751 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:36Z","lastTransitionTime":"2025-10-01T15:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.242405 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.258377 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.278737 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.293514 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.308956 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.320891 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.333880 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.338605 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.338644 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.338657 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.338673 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.338684 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:36Z","lastTransitionTime":"2025-10-01T15:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.348425 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.361862 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.380384 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.392239 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.402718 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.418263 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.432849 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.442684 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.442729 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.442740 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.442755 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.442765 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:36Z","lastTransitionTime":"2025-10-01T15:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.447597 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.462284 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.473662 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.490447 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.545348 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.545392 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.545404 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.545420 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.545435 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:36Z","lastTransitionTime":"2025-10-01T15:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.648761 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.648832 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.648845 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.648866 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.648879 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:36Z","lastTransitionTime":"2025-10-01T15:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.750923 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.750969 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.750984 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.751004 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.751016 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:36Z","lastTransitionTime":"2025-10-01T15:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.807526 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.807573 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:36 crc kubenswrapper[4726]: E1001 15:59:36.807780 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:36 crc kubenswrapper[4726]: E1001 15:59:36.807991 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.853318 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.853352 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.853361 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.853374 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.853383 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:36Z","lastTransitionTime":"2025-10-01T15:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.956120 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.956153 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.956169 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.956185 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:36 crc kubenswrapper[4726]: I1001 15:59:36.956196 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:36Z","lastTransitionTime":"2025-10-01T15:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.030625 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerStarted","Data":"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc"} Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.033806 4726 generic.go:334] "Generic (PLEG): container finished" podID="21c32caa-f2fd-4c2a-a766-52f3fc81a0ae" containerID="73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54" exitCode=0 Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.033891 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" event={"ID":"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae","Type":"ContainerDied","Data":"73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54"} Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.055875 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.059906 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.059957 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.059968 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.059981 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.059990 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:37Z","lastTransitionTime":"2025-10-01T15:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.073180 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.104243 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.124878 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.139900 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.152084 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.162201 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.162254 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.162268 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.162288 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.162307 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:37Z","lastTransitionTime":"2025-10-01T15:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.167999 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.179628 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.195030 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.207678 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.225222 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.241722 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.260485 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.264431 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.264465 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.264477 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.264494 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.264506 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:37Z","lastTransitionTime":"2025-10-01T15:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.276112 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.367114 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.367158 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.367170 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.367192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.367206 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:37Z","lastTransitionTime":"2025-10-01T15:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.469302 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.469342 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.469353 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.469367 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.469377 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:37Z","lastTransitionTime":"2025-10-01T15:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.470853 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:59:37 crc kubenswrapper[4726]: E1001 15:59:37.471005 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:59:45.47098635 +0000 UTC m=+38.372538927 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.471091 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:37 crc kubenswrapper[4726]: E1001 15:59:37.471210 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:59:37 crc kubenswrapper[4726]: E1001 15:59:37.471223 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:59:37 crc kubenswrapper[4726]: E1001 15:59:37.471233 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:37 crc kubenswrapper[4726]: E1001 15:59:37.471268 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:45.471261297 +0000 UTC m=+38.372813874 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.571451 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.571533 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.571551 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.571563 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.571611 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.571635 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.571574 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.571685 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:37Z","lastTransitionTime":"2025-10-01T15:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:37 crc kubenswrapper[4726]: E1001 15:59:37.571715 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:59:37 crc kubenswrapper[4726]: E1001 15:59:37.571752 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:45.571740451 +0000 UTC m=+38.473293028 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:59:37 crc kubenswrapper[4726]: E1001 15:59:37.571773 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:59:37 crc kubenswrapper[4726]: E1001 15:59:37.571819 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:59:37 crc kubenswrapper[4726]: E1001 15:59:37.571873 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:59:37 crc kubenswrapper[4726]: E1001 15:59:37.571899 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:37 crc kubenswrapper[4726]: E1001 15:59:37.571830 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:45.571814133 +0000 UTC m=+38.473366710 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:59:37 crc kubenswrapper[4726]: E1001 15:59:37.572005 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:45.571976298 +0000 UTC m=+38.473528915 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.674738 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.674775 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.674783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.674797 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.674806 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:37Z","lastTransitionTime":"2025-10-01T15:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.777880 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.777958 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.777976 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.777996 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.778012 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:37Z","lastTransitionTime":"2025-10-01T15:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.807369 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:37 crc kubenswrapper[4726]: E1001 15:59:37.807522 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.825636 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.847047 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.868542 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.879842 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.880929 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.880963 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.880974 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.880987 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.880997 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:37Z","lastTransitionTime":"2025-10-01T15:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.889941 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.915896 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.935322 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.946940 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.958598 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.972376 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.983648 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.983721 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.983738 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.983756 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.983769 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:37Z","lastTransitionTime":"2025-10-01T15:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:37 crc kubenswrapper[4726]: I1001 15:59:37.991451 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.003120 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.013477 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.028472 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.042716 4726 generic.go:334] "Generic (PLEG): container finished" podID="21c32caa-f2fd-4c2a-a766-52f3fc81a0ae" containerID="4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710" exitCode=0 Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.042800 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" event={"ID":"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae","Type":"ContainerDied","Data":"4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710"} Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.055317 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.073188 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.086272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.086310 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.086321 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.086341 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.086354 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:38Z","lastTransitionTime":"2025-10-01T15:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.088596 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.101507 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.116850 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.132289 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.156673 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.178204 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.188671 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.188770 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.188785 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.188803 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.188815 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:38Z","lastTransitionTime":"2025-10-01T15:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.193856 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.212973 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.229017 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.242471 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.255263 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.265683 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.290986 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.291026 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.291035 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.291066 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.291076 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:38Z","lastTransitionTime":"2025-10-01T15:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.394211 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.394252 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.394264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.394279 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.394290 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:38Z","lastTransitionTime":"2025-10-01T15:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.497157 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.497189 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.497198 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.497210 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.497220 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:38Z","lastTransitionTime":"2025-10-01T15:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.600127 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.600215 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.600237 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.600288 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.600308 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:38Z","lastTransitionTime":"2025-10-01T15:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.702401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.702435 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.702444 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.702457 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.702466 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:38Z","lastTransitionTime":"2025-10-01T15:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.805613 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.805735 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.805770 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.805908 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.805939 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:38Z","lastTransitionTime":"2025-10-01T15:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.807412 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.807483 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:38 crc kubenswrapper[4726]: E1001 15:59:38.807543 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:38 crc kubenswrapper[4726]: E1001 15:59:38.807669 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.909447 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.909517 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.909544 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.909577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:38 crc kubenswrapper[4726]: I1001 15:59:38.909604 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:38Z","lastTransitionTime":"2025-10-01T15:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.013250 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.013370 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.013400 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.013429 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.013450 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:39Z","lastTransitionTime":"2025-10-01T15:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.053455 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" event={"ID":"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae","Type":"ContainerStarted","Data":"c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0"} Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.067863 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerStarted","Data":"7f99b08bbd82bccbf1bad659961fb3b6d4f105ef046d430a6f41d5ea560fa880"} Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.068544 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.068603 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.073471 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.105692 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.109640 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.109900 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.116630 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.116710 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.116736 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.116769 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.116793 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:39Z","lastTransitionTime":"2025-10-01T15:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.126277 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.145213 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.164190 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.180905 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.191210 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.201918 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.211891 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.219540 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.219783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.219920 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.220020 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.220154 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:39Z","lastTransitionTime":"2025-10-01T15:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.228379 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.242884 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.255223 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.265361 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.275090 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.287684 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.302184 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.312146 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.323446 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.323478 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.323486 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.323498 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.323510 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:39Z","lastTransitionTime":"2025-10-01T15:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.328501 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.340343 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.357183 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.367528 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.378700 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.388668 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.410682 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f99b08bbd82bccbf1bad659961fb3b6d4f105ef046d430a6f41d5ea560fa880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.423538 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.425272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.425321 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.425335 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.425352 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.425363 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:39Z","lastTransitionTime":"2025-10-01T15:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.438844 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.451176 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.464104 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.528140 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.528196 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.528213 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.528228 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.528241 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:39Z","lastTransitionTime":"2025-10-01T15:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.630831 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.630875 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.630889 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.630907 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.630918 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:39Z","lastTransitionTime":"2025-10-01T15:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.732677 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.732717 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.732725 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.732739 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.732748 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:39Z","lastTransitionTime":"2025-10-01T15:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.808092 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:39 crc kubenswrapper[4726]: E1001 15:59:39.808242 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.835709 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.835781 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.835804 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.835835 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.835857 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:39Z","lastTransitionTime":"2025-10-01T15:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.938775 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.938821 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.938836 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.938857 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:39 crc kubenswrapper[4726]: I1001 15:59:39.938875 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:39Z","lastTransitionTime":"2025-10-01T15:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.042162 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.042225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.042241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.042262 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.042276 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:40Z","lastTransitionTime":"2025-10-01T15:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.071274 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.144941 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.145030 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.145101 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.145135 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.145159 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:40Z","lastTransitionTime":"2025-10-01T15:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.248458 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.248523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.248546 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.248577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.248600 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:40Z","lastTransitionTime":"2025-10-01T15:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.352172 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.352230 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.352248 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.352265 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.352277 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:40Z","lastTransitionTime":"2025-10-01T15:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.455067 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.455114 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.455126 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.455142 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.455154 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:40Z","lastTransitionTime":"2025-10-01T15:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.557594 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.557671 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.557696 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.557724 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.557746 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:40Z","lastTransitionTime":"2025-10-01T15:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.660677 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.660725 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.660740 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.660763 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.660778 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:40Z","lastTransitionTime":"2025-10-01T15:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.763589 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.763638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.763655 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.763676 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.763690 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:40Z","lastTransitionTime":"2025-10-01T15:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.807297 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.807342 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:40 crc kubenswrapper[4726]: E1001 15:59:40.807432 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:40 crc kubenswrapper[4726]: E1001 15:59:40.807593 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.866607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.866669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.866681 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.866715 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.866726 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:40Z","lastTransitionTime":"2025-10-01T15:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.968964 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.969067 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.969080 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.969095 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:40 crc kubenswrapper[4726]: I1001 15:59:40.969105 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:40Z","lastTransitionTime":"2025-10-01T15:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.039526 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.039562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.039571 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.039615 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.039642 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:41Z","lastTransitionTime":"2025-10-01T15:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:41 crc kubenswrapper[4726]: E1001 15:59:41.051543 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.055177 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.055216 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.055225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.055267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.055278 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:41Z","lastTransitionTime":"2025-10-01T15:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:41 crc kubenswrapper[4726]: E1001 15:59:41.067119 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.070596 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.070657 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.070678 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.070700 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.070718 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:41Z","lastTransitionTime":"2025-10-01T15:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.073599 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 15:59:41 crc kubenswrapper[4726]: E1001 15:59:41.083495 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.086738 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.086766 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.086781 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.086796 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.086805 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:41Z","lastTransitionTime":"2025-10-01T15:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:41 crc kubenswrapper[4726]: E1001 15:59:41.101528 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.105268 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.105309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.105323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.105340 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.105349 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:41Z","lastTransitionTime":"2025-10-01T15:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:41 crc kubenswrapper[4726]: E1001 15:59:41.118192 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:41 crc kubenswrapper[4726]: E1001 15:59:41.118323 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.119834 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.119881 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.119890 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.119906 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.119916 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:41Z","lastTransitionTime":"2025-10-01T15:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.222085 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.222142 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.222160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.222183 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.222200 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:41Z","lastTransitionTime":"2025-10-01T15:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.324840 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.324893 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.324912 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.324934 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.324951 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:41Z","lastTransitionTime":"2025-10-01T15:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.427442 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.427500 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.427509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.427523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.427532 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:41Z","lastTransitionTime":"2025-10-01T15:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.530257 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.530290 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.530299 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.530311 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.530321 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:41Z","lastTransitionTime":"2025-10-01T15:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.633513 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.634009 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.634295 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.634489 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.634629 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:41Z","lastTransitionTime":"2025-10-01T15:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.738378 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.738452 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.738477 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.738510 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.738534 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:41Z","lastTransitionTime":"2025-10-01T15:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.807283 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:41 crc kubenswrapper[4726]: E1001 15:59:41.808462 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.841411 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.841659 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.841766 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.841838 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.841894 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:41Z","lastTransitionTime":"2025-10-01T15:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.944613 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.944939 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.945092 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.945181 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:41 crc kubenswrapper[4726]: I1001 15:59:41.945265 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:41Z","lastTransitionTime":"2025-10-01T15:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.047963 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.048022 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.048041 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.048084 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.048101 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:42Z","lastTransitionTime":"2025-10-01T15:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.078817 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/0.log" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.082390 4726 generic.go:334] "Generic (PLEG): container finished" podID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerID="7f99b08bbd82bccbf1bad659961fb3b6d4f105ef046d430a6f41d5ea560fa880" exitCode=1 Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.082469 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerDied","Data":"7f99b08bbd82bccbf1bad659961fb3b6d4f105ef046d430a6f41d5ea560fa880"} Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.083902 4726 scope.go:117] "RemoveContainer" containerID="7f99b08bbd82bccbf1bad659961fb3b6d4f105ef046d430a6f41d5ea560fa880" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.099843 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:42Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.113322 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:42Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.132872 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:42Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.150485 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:42Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.151006 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.151195 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.151225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.151247 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.151262 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:42Z","lastTransitionTime":"2025-10-01T15:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.169839 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:42Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.182068 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:42Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.192273 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:42Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.204848 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:42Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.227942 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:42Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.253591 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.253638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.253650 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.253669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.253683 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:42Z","lastTransitionTime":"2025-10-01T15:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.272982 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:42Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.300080 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f99b08bbd82bccbf1bad659961fb3b6d4f105ef046d430a6f41d5ea560fa880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f99b08bbd82bccbf1bad659961fb3b6d4f105ef046d430a6f41d5ea560fa880\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:42Z\\\",\\\"message\\\":\\\"5:59:41.526635 6034 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 15:59:41.527710 6034 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 15:59:41.527737 6034 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 15:59:41.529939 6034 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 15:59:41.529976 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 15:59:41.530079 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 15:59:41.530138 6034 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 15:59:41.530156 6034 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 15:59:41.530196 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 15:59:41.530222 6034 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 15:59:41.530278 6034 factory.go:656] Stopping watch factory\\\\nI1001 15:59:41.530891 6034 ovnkube.go:599] Stopped ovnkube\\\\nI1001 15:59:41.530668 6034 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 15:59:41.530679 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 15:59:41.530693 6034 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 15:59:41.530957 6034 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 15:59:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:42Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.317131 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:42Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.334526 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:42Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.346705 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:42Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.355508 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.355541 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.355552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.355567 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.355580 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:42Z","lastTransitionTime":"2025-10-01T15:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.458723 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.458779 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.458791 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.458811 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.458822 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:42Z","lastTransitionTime":"2025-10-01T15:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.562285 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.562755 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.563217 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.563460 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.563632 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:42Z","lastTransitionTime":"2025-10-01T15:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.665872 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.666272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.666655 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.667091 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.667266 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:42Z","lastTransitionTime":"2025-10-01T15:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.770254 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.770531 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.770644 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.770733 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.770809 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:42Z","lastTransitionTime":"2025-10-01T15:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.807847 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:42 crc kubenswrapper[4726]: E1001 15:59:42.807975 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.808601 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:42 crc kubenswrapper[4726]: E1001 15:59:42.808834 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.873280 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.873328 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.873338 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.873354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.873365 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:42Z","lastTransitionTime":"2025-10-01T15:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.976045 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.976103 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.976113 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.976134 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:42 crc kubenswrapper[4726]: I1001 15:59:42.976144 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:42Z","lastTransitionTime":"2025-10-01T15:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.078134 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.078174 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.078184 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.078198 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.078217 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:43Z","lastTransitionTime":"2025-10-01T15:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.086787 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/0.log" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.089684 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerStarted","Data":"7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf"} Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.089807 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.104995 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.136393 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.167029 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.180579 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.181752 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.181810 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.181829 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.181855 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.181876 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:43Z","lastTransitionTime":"2025-10-01T15:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.193252 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.210159 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.227310 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.240006 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.254216 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.268514 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.282934 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.283613 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.283648 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.283661 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.283675 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.283686 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:43Z","lastTransitionTime":"2025-10-01T15:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.301203 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f99b08bbd82bccbf1bad659961fb3b6d4f105ef046d430a6f41d5ea560fa880\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:42Z\\\",\\\"message\\\":\\\"5:59:41.526635 6034 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 15:59:41.527710 6034 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 15:59:41.527737 6034 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 15:59:41.529939 6034 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 15:59:41.529976 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 15:59:41.530079 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 15:59:41.530138 6034 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 15:59:41.530156 6034 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 15:59:41.530196 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 15:59:41.530222 6034 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 15:59:41.530278 6034 factory.go:656] Stopping watch factory\\\\nI1001 15:59:41.530891 6034 ovnkube.go:599] Stopped ovnkube\\\\nI1001 15:59:41.530668 6034 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 15:59:41.530679 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 15:59:41.530693 6034 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 15:59:41.530957 6034 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 15:59:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.315268 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.327871 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.385732 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.385779 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.385790 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.385806 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.385816 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:43Z","lastTransitionTime":"2025-10-01T15:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.487799 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.487835 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.487846 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.487864 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.487876 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:43Z","lastTransitionTime":"2025-10-01T15:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.590584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.590628 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.590640 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.590656 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.590668 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:43Z","lastTransitionTime":"2025-10-01T15:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.629011 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl"] Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.630902 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.633472 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.633883 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.646753 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.663066 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.682722 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.692541 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.692700 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.692765 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.692827 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.692883 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:43Z","lastTransitionTime":"2025-10-01T15:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.694946 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.705737 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.718366 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.735449 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.736783 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a8be9689-98f6-45f3-9523-3b66f3a55c1e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vdljl\" (UID: \"a8be9689-98f6-45f3-9523-3b66f3a55c1e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.736918 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a8be9689-98f6-45f3-9523-3b66f3a55c1e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vdljl\" (UID: \"a8be9689-98f6-45f3-9523-3b66f3a55c1e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.737035 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwhdc\" (UniqueName: \"kubernetes.io/projected/a8be9689-98f6-45f3-9523-3b66f3a55c1e-kube-api-access-kwhdc\") pod \"ovnkube-control-plane-749d76644c-vdljl\" (UID: \"a8be9689-98f6-45f3-9523-3b66f3a55c1e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.737159 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a8be9689-98f6-45f3-9523-3b66f3a55c1e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vdljl\" (UID: \"a8be9689-98f6-45f3-9523-3b66f3a55c1e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.746153 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.757231 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.774948 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.789378 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.794662 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.794730 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.794741 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.794754 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.794763 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:43Z","lastTransitionTime":"2025-10-01T15:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.802401 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.808012 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:43 crc kubenswrapper[4726]: E1001 15:59:43.808134 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.812752 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.829249 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f99b08bbd82bccbf1bad659961fb3b6d4f105ef046d430a6f41d5ea560fa880\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:42Z\\\",\\\"message\\\":\\\"5:59:41.526635 6034 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 15:59:41.527710 6034 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 15:59:41.527737 6034 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 15:59:41.529939 6034 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 15:59:41.529976 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 15:59:41.530079 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 15:59:41.530138 6034 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 15:59:41.530156 6034 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 15:59:41.530196 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 15:59:41.530222 6034 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 15:59:41.530278 6034 factory.go:656] Stopping watch factory\\\\nI1001 15:59:41.530891 6034 ovnkube.go:599] Stopped ovnkube\\\\nI1001 15:59:41.530668 6034 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 15:59:41.530679 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 15:59:41.530693 6034 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 15:59:41.530957 6034 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 15:59:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.838497 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a8be9689-98f6-45f3-9523-3b66f3a55c1e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vdljl\" (UID: \"a8be9689-98f6-45f3-9523-3b66f3a55c1e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.838583 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a8be9689-98f6-45f3-9523-3b66f3a55c1e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vdljl\" (UID: \"a8be9689-98f6-45f3-9523-3b66f3a55c1e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.838620 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwhdc\" (UniqueName: \"kubernetes.io/projected/a8be9689-98f6-45f3-9523-3b66f3a55c1e-kube-api-access-kwhdc\") pod \"ovnkube-control-plane-749d76644c-vdljl\" (UID: \"a8be9689-98f6-45f3-9523-3b66f3a55c1e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.838643 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a8be9689-98f6-45f3-9523-3b66f3a55c1e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vdljl\" (UID: \"a8be9689-98f6-45f3-9523-3b66f3a55c1e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.839149 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a8be9689-98f6-45f3-9523-3b66f3a55c1e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vdljl\" (UID: \"a8be9689-98f6-45f3-9523-3b66f3a55c1e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.839295 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a8be9689-98f6-45f3-9523-3b66f3a55c1e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vdljl\" (UID: \"a8be9689-98f6-45f3-9523-3b66f3a55c1e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.841346 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.847393 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a8be9689-98f6-45f3-9523-3b66f3a55c1e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vdljl\" (UID: \"a8be9689-98f6-45f3-9523-3b66f3a55c1e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.854736 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwhdc\" (UniqueName: \"kubernetes.io/projected/a8be9689-98f6-45f3-9523-3b66f3a55c1e-kube-api-access-kwhdc\") pod \"ovnkube-control-plane-749d76644c-vdljl\" (UID: \"a8be9689-98f6-45f3-9523-3b66f3a55c1e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.897241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.897278 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.897286 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.897300 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.897308 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:43Z","lastTransitionTime":"2025-10-01T15:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:43 crc kubenswrapper[4726]: I1001 15:59:43.943609 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" Oct 01 15:59:43 crc kubenswrapper[4726]: W1001 15:59:43.962844 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8be9689_98f6_45f3_9523_3b66f3a55c1e.slice/crio-5a86392d720e858715a9d8087097b6ef2a625c3449fe64778d290d2b4789813d WatchSource:0}: Error finding container 5a86392d720e858715a9d8087097b6ef2a625c3449fe64778d290d2b4789813d: Status 404 returned error can't find the container with id 5a86392d720e858715a9d8087097b6ef2a625c3449fe64778d290d2b4789813d Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.000177 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.000221 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.000233 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.000251 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.000263 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:44Z","lastTransitionTime":"2025-10-01T15:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.097203 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/1.log" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.098003 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/0.log" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.103022 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.103084 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.103102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.103122 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.103136 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:44Z","lastTransitionTime":"2025-10-01T15:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.103146 4726 generic.go:334] "Generic (PLEG): container finished" podID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerID="7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf" exitCode=1 Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.103219 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerDied","Data":"7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf"} Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.103298 4726 scope.go:117] "RemoveContainer" containerID="7f99b08bbd82bccbf1bad659961fb3b6d4f105ef046d430a6f41d5ea560fa880" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.104547 4726 scope.go:117] "RemoveContainer" containerID="7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf" Oct 01 15:59:44 crc kubenswrapper[4726]: E1001 15:59:44.104827 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.105522 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" event={"ID":"a8be9689-98f6-45f3-9523-3b66f3a55c1e","Type":"ContainerStarted","Data":"5a86392d720e858715a9d8087097b6ef2a625c3449fe64778d290d2b4789813d"} Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.119608 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.134404 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.149608 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.161074 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.171925 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.184675 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.196175 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.205854 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.205897 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.205918 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.205936 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.205946 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:44Z","lastTransitionTime":"2025-10-01T15:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.206812 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.217202 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.229728 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.248232 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f99b08bbd82bccbf1bad659961fb3b6d4f105ef046d430a6f41d5ea560fa880\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:42Z\\\",\\\"message\\\":\\\"5:59:41.526635 6034 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 15:59:41.527710 6034 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 15:59:41.527737 6034 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 15:59:41.529939 6034 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 15:59:41.529976 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 15:59:41.530079 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 15:59:41.530138 6034 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 15:59:41.530156 6034 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 15:59:41.530196 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 15:59:41.530222 6034 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 15:59:41.530278 6034 factory.go:656] Stopping watch factory\\\\nI1001 15:59:41.530891 6034 ovnkube.go:599] Stopped ovnkube\\\\nI1001 15:59:41.530668 6034 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 15:59:41.530679 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 15:59:41.530693 6034 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 15:59:41.530957 6034 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 15:59:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"message\\\":\\\"169],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:59:43.163371 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:59:43.163377 6156 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1001 15:59:43.163391 6156 services_controller.go:443] Built service openshift-kube-scheduler/scheduler LB cluster-wide conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.261719 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.273679 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.284515 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.294678 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.308264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.308298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.308307 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.308319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.308328 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:44Z","lastTransitionTime":"2025-10-01T15:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.412002 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.412074 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.412089 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.412108 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.412119 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:44Z","lastTransitionTime":"2025-10-01T15:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.515367 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.515406 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.515418 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.515435 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.515446 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:44Z","lastTransitionTime":"2025-10-01T15:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.618323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.618366 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.618377 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.618392 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.618404 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:44Z","lastTransitionTime":"2025-10-01T15:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.721503 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.721555 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.721566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.721584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.721599 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:44Z","lastTransitionTime":"2025-10-01T15:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.807373 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.807434 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:44 crc kubenswrapper[4726]: E1001 15:59:44.807520 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:44 crc kubenswrapper[4726]: E1001 15:59:44.807648 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.824436 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.824528 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.824560 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.824597 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.824623 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:44Z","lastTransitionTime":"2025-10-01T15:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.928141 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.928192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.928207 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.928227 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:44 crc kubenswrapper[4726]: I1001 15:59:44.928244 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:44Z","lastTransitionTime":"2025-10-01T15:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.031556 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.031641 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.031667 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.031698 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.031750 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:45Z","lastTransitionTime":"2025-10-01T15:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.111553 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" event={"ID":"a8be9689-98f6-45f3-9523-3b66f3a55c1e","Type":"ContainerStarted","Data":"817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2"} Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.111610 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" event={"ID":"a8be9689-98f6-45f3-9523-3b66f3a55c1e","Type":"ContainerStarted","Data":"dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da"} Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.113627 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/1.log" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.133219 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.136750 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.136792 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.136805 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.136823 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.136837 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:45Z","lastTransitionTime":"2025-10-01T15:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.153022 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.180264 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f99b08bbd82bccbf1bad659961fb3b6d4f105ef046d430a6f41d5ea560fa880\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:42Z\\\",\\\"message\\\":\\\"5:59:41.526635 6034 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 15:59:41.527710 6034 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 15:59:41.527737 6034 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 15:59:41.529939 6034 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 15:59:41.529976 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 15:59:41.530079 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 15:59:41.530138 6034 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 15:59:41.530156 6034 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 15:59:41.530196 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 15:59:41.530222 6034 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 15:59:41.530278 6034 factory.go:656] Stopping watch factory\\\\nI1001 15:59:41.530891 6034 ovnkube.go:599] Stopped ovnkube\\\\nI1001 15:59:41.530668 6034 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 15:59:41.530679 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 15:59:41.530693 6034 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 15:59:41.530957 6034 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 15:59:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"message\\\":\\\"169],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:59:43.163371 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:59:43.163377 6156 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1001 15:59:43.163391 6156 services_controller.go:443] Built service openshift-kube-scheduler/scheduler LB cluster-wide conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.203515 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.221236 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.238277 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.238880 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.238935 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.238953 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.238978 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.238998 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:45Z","lastTransitionTime":"2025-10-01T15:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.263490 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.280895 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.300476 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.319960 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.341746 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.341784 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.341794 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.341809 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.341819 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:45Z","lastTransitionTime":"2025-10-01T15:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.345421 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.361701 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.376210 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.392808 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.408017 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.444807 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.444880 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.444904 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.444933 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.444960 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:45Z","lastTransitionTime":"2025-10-01T15:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.478266 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-4pzd8"] Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.478837 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.478919 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.492741 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.506240 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.526888 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.541013 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.547435 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.547488 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.547502 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.547518 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.547528 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:45Z","lastTransitionTime":"2025-10-01T15:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.559062 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.559167 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:00:01.559146893 +0000 UTC m=+54.460699470 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.559246 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.559286 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs\") pod \"network-metrics-daemon-4pzd8\" (UID: \"2a22e7a8-726a-4538-a27d-44c58f89f0b1\") " pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.559304 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gccgk\" (UniqueName: \"kubernetes.io/projected/2a22e7a8-726a-4538-a27d-44c58f89f0b1-kube-api-access-gccgk\") pod \"network-metrics-daemon-4pzd8\" (UID: \"2a22e7a8-726a-4538-a27d-44c58f89f0b1\") " pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.559444 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.559464 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.559476 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.559523 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 16:00:01.559512922 +0000 UTC m=+54.461065589 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.563398 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f99b08bbd82bccbf1bad659961fb3b6d4f105ef046d430a6f41d5ea560fa880\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:42Z\\\",\\\"message\\\":\\\"5:59:41.526635 6034 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 15:59:41.527710 6034 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 15:59:41.527737 6034 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 15:59:41.529939 6034 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 15:59:41.529976 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 15:59:41.530079 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 15:59:41.530138 6034 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 15:59:41.530156 6034 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 15:59:41.530196 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 15:59:41.530222 6034 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 15:59:41.530278 6034 factory.go:656] Stopping watch factory\\\\nI1001 15:59:41.530891 6034 ovnkube.go:599] Stopped ovnkube\\\\nI1001 15:59:41.530668 6034 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 15:59:41.530679 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 15:59:41.530693 6034 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 15:59:41.530957 6034 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 15:59:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"message\\\":\\\"169],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:59:43.163371 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:59:43.163377 6156 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1001 15:59:43.163391 6156 services_controller.go:443] Built service openshift-kube-scheduler/scheduler LB cluster-wide conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.578095 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.599247 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.611157 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.622915 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.635381 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.650695 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.650941 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.650957 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.650978 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.651011 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:45Z","lastTransitionTime":"2025-10-01T15:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.652654 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.660186 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gccgk\" (UniqueName: \"kubernetes.io/projected/2a22e7a8-726a-4538-a27d-44c58f89f0b1-kube-api-access-gccgk\") pod \"network-metrics-daemon-4pzd8\" (UID: \"2a22e7a8-726a-4538-a27d-44c58f89f0b1\") " pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.660247 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.660286 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.660340 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.660379 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs\") pod \"network-metrics-daemon-4pzd8\" (UID: \"2a22e7a8-726a-4538-a27d-44c58f89f0b1\") " pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.660402 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.660472 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 16:00:01.660450797 +0000 UTC m=+54.562003394 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.660477 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.660527 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs podName:2a22e7a8-726a-4538-a27d-44c58f89f0b1 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:46.160509149 +0000 UTC m=+39.062061746 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs") pod "network-metrics-daemon-4pzd8" (UID: "2a22e7a8-726a-4538-a27d-44c58f89f0b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.660547 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.660585 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.660601 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.660676 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 16:00:01.660655233 +0000 UTC m=+54.562207870 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.660738 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.660781 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 16:00:01.660770856 +0000 UTC m=+54.562323503 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.666253 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.681068 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gccgk\" (UniqueName: \"kubernetes.io/projected/2a22e7a8-726a-4538-a27d-44c58f89f0b1-kube-api-access-gccgk\") pod \"network-metrics-daemon-4pzd8\" (UID: \"2a22e7a8-726a-4538-a27d-44c58f89f0b1\") " pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.682286 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.694760 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.705701 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.722477 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:45Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.753624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.753663 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.753674 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.753693 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.753705 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:45Z","lastTransitionTime":"2025-10-01T15:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.807892 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:45 crc kubenswrapper[4726]: E1001 15:59:45.808103 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.855873 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.856519 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.856659 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.856781 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.856901 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:45Z","lastTransitionTime":"2025-10-01T15:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.960146 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.960487 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.960690 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.960927 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:45 crc kubenswrapper[4726]: I1001 15:59:45.961160 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:45Z","lastTransitionTime":"2025-10-01T15:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.064027 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.064093 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.064112 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.064130 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.064148 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:46Z","lastTransitionTime":"2025-10-01T15:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.166375 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs\") pod \"network-metrics-daemon-4pzd8\" (UID: \"2a22e7a8-726a-4538-a27d-44c58f89f0b1\") " pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:46 crc kubenswrapper[4726]: E1001 15:59:46.166577 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:59:46 crc kubenswrapper[4726]: E1001 15:59:46.166655 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs podName:2a22e7a8-726a-4538-a27d-44c58f89f0b1 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:47.166628363 +0000 UTC m=+40.068180970 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs") pod "network-metrics-daemon-4pzd8" (UID: "2a22e7a8-726a-4538-a27d-44c58f89f0b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.167470 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.167545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.167566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.167593 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.167615 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:46Z","lastTransitionTime":"2025-10-01T15:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.269394 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.269749 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.269761 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.269782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.269794 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:46Z","lastTransitionTime":"2025-10-01T15:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.371889 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.372145 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.372229 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.372307 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.372417 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:46Z","lastTransitionTime":"2025-10-01T15:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.475013 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.475092 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.475114 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.475135 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.475149 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:46Z","lastTransitionTime":"2025-10-01T15:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.578481 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.578538 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.578549 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.578575 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.578599 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:46Z","lastTransitionTime":"2025-10-01T15:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.611489 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.612524 4726 scope.go:117] "RemoveContainer" containerID="7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf" Oct 01 15:59:46 crc kubenswrapper[4726]: E1001 15:59:46.612717 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.632700 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.646617 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.664142 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.679524 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.681226 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.681258 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.681271 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.682137 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.682209 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:46Z","lastTransitionTime":"2025-10-01T15:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.691989 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.705846 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.720141 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.734797 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.748564 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.761409 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.785084 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.785127 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.785146 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.785164 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.785176 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:46Z","lastTransitionTime":"2025-10-01T15:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.785409 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"message\\\":\\\"169],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:59:43.163371 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:59:43.163377 6156 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1001 15:59:43.163391 6156 services_controller.go:443] Built service openshift-kube-scheduler/scheduler LB cluster-wide conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.800775 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.807150 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:46 crc kubenswrapper[4726]: E1001 15:59:46.807471 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.807279 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:46 crc kubenswrapper[4726]: E1001 15:59:46.807745 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.807225 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:46 crc kubenswrapper[4726]: E1001 15:59:46.808010 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.813898 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.827112 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.838570 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.854289 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:46Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.888085 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.888709 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.888737 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.888757 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.888771 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:46Z","lastTransitionTime":"2025-10-01T15:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.990823 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.991145 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.991415 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.991508 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:46 crc kubenswrapper[4726]: I1001 15:59:46.991580 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:46Z","lastTransitionTime":"2025-10-01T15:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.094471 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.094749 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.095039 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.095239 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.095374 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:47Z","lastTransitionTime":"2025-10-01T15:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.177910 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs\") pod \"network-metrics-daemon-4pzd8\" (UID: \"2a22e7a8-726a-4538-a27d-44c58f89f0b1\") " pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:47 crc kubenswrapper[4726]: E1001 15:59:47.178068 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:59:47 crc kubenswrapper[4726]: E1001 15:59:47.178129 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs podName:2a22e7a8-726a-4538-a27d-44c58f89f0b1 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:49.178112162 +0000 UTC m=+42.079664739 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs") pod "network-metrics-daemon-4pzd8" (UID: "2a22e7a8-726a-4538-a27d-44c58f89f0b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.203482 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.203712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.203806 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.203890 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.203969 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:47Z","lastTransitionTime":"2025-10-01T15:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.306264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.306292 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.306300 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.306315 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.306323 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:47Z","lastTransitionTime":"2025-10-01T15:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.409184 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.409258 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.409268 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.409282 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.409294 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:47Z","lastTransitionTime":"2025-10-01T15:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.511342 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.511382 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.511392 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.511406 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.511415 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:47Z","lastTransitionTime":"2025-10-01T15:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.613396 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.613435 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.613447 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.613463 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.613474 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:47Z","lastTransitionTime":"2025-10-01T15:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.715794 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.715864 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.715886 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.715909 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.715926 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:47Z","lastTransitionTime":"2025-10-01T15:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.807977 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:47 crc kubenswrapper[4726]: E1001 15:59:47.808315 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.808493 4726 scope.go:117] "RemoveContainer" containerID="1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.817903 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.817948 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.817963 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.818000 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.818018 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:47Z","lastTransitionTime":"2025-10-01T15:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.832830 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"message\\\":\\\"169],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:59:43.163371 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:59:43.163377 6156 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1001 15:59:43.163391 6156 services_controller.go:443] Built service openshift-kube-scheduler/scheduler LB cluster-wide conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:47Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.849638 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:47Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.863471 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:47Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.875696 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:47Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.889230 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:47Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.900199 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:47Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.912345 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:47Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.920099 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.920132 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.920141 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.920153 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.920163 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:47Z","lastTransitionTime":"2025-10-01T15:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.924956 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:47Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.941674 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:47Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.951389 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:47Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.964804 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:47Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.979412 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:47Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:47 crc kubenswrapper[4726]: I1001 15:59:47.994165 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:47Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.007970 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.018738 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.022519 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.022592 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.022611 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.022635 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.022652 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:48Z","lastTransitionTime":"2025-10-01T15:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.031511 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.124131 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.124158 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.124166 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.124179 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.124188 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:48Z","lastTransitionTime":"2025-10-01T15:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.125831 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.127327 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb"} Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.128028 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.144824 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.158980 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.172949 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.185322 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.207278 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"message\\\":\\\"169],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:59:43.163371 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:59:43.163377 6156 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1001 15:59:43.163391 6156 services_controller.go:443] Built service openshift-kube-scheduler/scheduler LB cluster-wide conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.224212 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.226103 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.226137 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.226148 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.226164 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.226176 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:48Z","lastTransitionTime":"2025-10-01T15:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.240226 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.250834 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.266923 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.278339 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.291004 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.303669 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.314757 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.328006 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.328506 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.328541 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.328552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.328568 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.328578 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:48Z","lastTransitionTime":"2025-10-01T15:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.339651 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.354120 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:48Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.431273 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.431324 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.431337 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.431355 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.431369 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:48Z","lastTransitionTime":"2025-10-01T15:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.534112 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.534149 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.534158 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.534171 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.534180 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:48Z","lastTransitionTime":"2025-10-01T15:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.635883 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.635918 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.635928 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.635941 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.635949 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:48Z","lastTransitionTime":"2025-10-01T15:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.738554 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.738603 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.738617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.738634 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.738646 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:48Z","lastTransitionTime":"2025-10-01T15:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.807564 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:48 crc kubenswrapper[4726]: E1001 15:59:48.807911 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.807914 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:48 crc kubenswrapper[4726]: E1001 15:59:48.808014 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.807952 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:48 crc kubenswrapper[4726]: E1001 15:59:48.808110 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.840743 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.840790 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.840801 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.840853 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.840867 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:48Z","lastTransitionTime":"2025-10-01T15:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.943788 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.943860 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.943874 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.943894 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:48 crc kubenswrapper[4726]: I1001 15:59:48.943906 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:48Z","lastTransitionTime":"2025-10-01T15:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.046068 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.046119 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.046133 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.046151 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.046166 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:49Z","lastTransitionTime":"2025-10-01T15:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.148619 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.148675 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.148688 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.148704 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.148717 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:49Z","lastTransitionTime":"2025-10-01T15:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.198072 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs\") pod \"network-metrics-daemon-4pzd8\" (UID: \"2a22e7a8-726a-4538-a27d-44c58f89f0b1\") " pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:49 crc kubenswrapper[4726]: E1001 15:59:49.198267 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:59:49 crc kubenswrapper[4726]: E1001 15:59:49.198348 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs podName:2a22e7a8-726a-4538-a27d-44c58f89f0b1 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:53.198330618 +0000 UTC m=+46.099883195 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs") pod "network-metrics-daemon-4pzd8" (UID: "2a22e7a8-726a-4538-a27d-44c58f89f0b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.251612 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.251647 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.251656 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.251667 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.251676 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:49Z","lastTransitionTime":"2025-10-01T15:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.354690 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.354777 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.354847 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.354865 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.354877 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:49Z","lastTransitionTime":"2025-10-01T15:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.457983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.458105 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.458150 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.458212 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.458239 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:49Z","lastTransitionTime":"2025-10-01T15:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.561694 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.561743 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.561759 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.561779 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.561794 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:49Z","lastTransitionTime":"2025-10-01T15:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.665384 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.665462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.665485 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.665502 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.665513 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:49Z","lastTransitionTime":"2025-10-01T15:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.767841 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.767938 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.767966 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.767992 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.768012 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:49Z","lastTransitionTime":"2025-10-01T15:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.807618 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:49 crc kubenswrapper[4726]: E1001 15:59:49.807876 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.869890 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.869939 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.869955 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.869972 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.869982 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:49Z","lastTransitionTime":"2025-10-01T15:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.972445 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.972513 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.972524 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.972543 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:49 crc kubenswrapper[4726]: I1001 15:59:49.972555 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:49Z","lastTransitionTime":"2025-10-01T15:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.075531 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.075572 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.075582 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.075596 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.075606 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:50Z","lastTransitionTime":"2025-10-01T15:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.178155 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.178385 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.178427 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.178457 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.178481 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:50Z","lastTransitionTime":"2025-10-01T15:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.281475 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.281524 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.281536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.281553 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.281564 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:50Z","lastTransitionTime":"2025-10-01T15:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.384883 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.384966 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.384982 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.385006 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.385023 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:50Z","lastTransitionTime":"2025-10-01T15:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.487499 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.487573 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.487591 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.487616 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.487634 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:50Z","lastTransitionTime":"2025-10-01T15:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.591157 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.591229 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.591240 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.591261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.591277 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:50Z","lastTransitionTime":"2025-10-01T15:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.694065 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.694273 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.694288 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.694309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.694321 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:50Z","lastTransitionTime":"2025-10-01T15:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.798881 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.798949 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.798967 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.798990 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.799006 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:50Z","lastTransitionTime":"2025-10-01T15:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.807237 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.807377 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.807620 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:50 crc kubenswrapper[4726]: E1001 15:59:50.807607 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 15:59:50 crc kubenswrapper[4726]: E1001 15:59:50.807753 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:50 crc kubenswrapper[4726]: E1001 15:59:50.807904 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.902278 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.902328 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.902342 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.902362 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:50 crc kubenswrapper[4726]: I1001 15:59:50.902376 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:50Z","lastTransitionTime":"2025-10-01T15:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.006484 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.006549 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.006565 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.006587 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.006599 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.109959 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.109994 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.110006 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.110021 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.110040 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.211810 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.211851 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.211863 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.211880 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.211892 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.265556 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.265718 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.265732 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.265751 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.265762 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: E1001 15:59:51.279654 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.284977 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.285033 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.285065 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.285091 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.285103 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: E1001 15:59:51.296796 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.300856 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.300924 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.300937 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.300957 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.300968 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: E1001 15:59:51.315314 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.318759 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.318806 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.318816 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.318836 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.318847 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: E1001 15:59:51.333462 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.338966 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.339033 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.339065 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.339102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.339122 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: E1001 15:59:51.356743 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:51 crc kubenswrapper[4726]: E1001 15:59:51.356993 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.358979 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.359019 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.359031 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.359070 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.359093 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.461239 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.461295 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.461307 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.461326 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.461338 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.564031 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.564104 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.564115 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.564133 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.564144 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.667334 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.667411 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.667432 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.667455 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.667473 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.770405 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.770459 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.770483 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.770511 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.770532 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.807947 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:51 crc kubenswrapper[4726]: E1001 15:59:51.808146 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.872638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.872671 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.872683 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.872700 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.872712 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.974754 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.975069 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.975173 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.975273 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:51 crc kubenswrapper[4726]: I1001 15:59:51.975454 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:51Z","lastTransitionTime":"2025-10-01T15:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.077441 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.077497 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.077514 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.077536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.077554 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:52Z","lastTransitionTime":"2025-10-01T15:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.180019 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.180062 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.180071 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.180084 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.180092 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:52Z","lastTransitionTime":"2025-10-01T15:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.282822 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.283118 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.283220 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.283315 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.283406 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:52Z","lastTransitionTime":"2025-10-01T15:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.385503 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.385559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.385577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.385604 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.385629 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:52Z","lastTransitionTime":"2025-10-01T15:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.488911 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.488958 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.488976 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.488998 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.489015 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:52Z","lastTransitionTime":"2025-10-01T15:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.591186 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.591250 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.591272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.591292 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.591308 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:52Z","lastTransitionTime":"2025-10-01T15:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.694163 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.694213 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.694225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.694239 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.694249 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:52Z","lastTransitionTime":"2025-10-01T15:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.797012 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.797090 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.797106 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.797124 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.797153 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:52Z","lastTransitionTime":"2025-10-01T15:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.807201 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.807287 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:52 crc kubenswrapper[4726]: E1001 15:59:52.807329 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:52 crc kubenswrapper[4726]: E1001 15:59:52.807414 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.807294 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:52 crc kubenswrapper[4726]: E1001 15:59:52.807500 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.899208 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.899266 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.899280 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.899298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:52 crc kubenswrapper[4726]: I1001 15:59:52.899310 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:52Z","lastTransitionTime":"2025-10-01T15:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.001980 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.002033 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.002042 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.002069 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.002081 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:53Z","lastTransitionTime":"2025-10-01T15:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.104803 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.104854 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.104866 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.104884 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.104898 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:53Z","lastTransitionTime":"2025-10-01T15:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.207717 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.207762 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.207773 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.207792 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.207805 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:53Z","lastTransitionTime":"2025-10-01T15:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.242021 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs\") pod \"network-metrics-daemon-4pzd8\" (UID: \"2a22e7a8-726a-4538-a27d-44c58f89f0b1\") " pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:53 crc kubenswrapper[4726]: E1001 15:59:53.242276 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:59:53 crc kubenswrapper[4726]: E1001 15:59:53.242342 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs podName:2a22e7a8-726a-4538-a27d-44c58f89f0b1 nodeName:}" failed. No retries permitted until 2025-10-01 16:00:01.242321541 +0000 UTC m=+54.143874128 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs") pod "network-metrics-daemon-4pzd8" (UID: "2a22e7a8-726a-4538-a27d-44c58f89f0b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.310202 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.310261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.310278 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.310304 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.310322 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:53Z","lastTransitionTime":"2025-10-01T15:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.413017 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.413124 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.413151 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.413174 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.413192 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:53Z","lastTransitionTime":"2025-10-01T15:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.515498 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.515543 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.515559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.515578 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.515591 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:53Z","lastTransitionTime":"2025-10-01T15:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.619154 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.619210 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.619229 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.619256 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.619274 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:53Z","lastTransitionTime":"2025-10-01T15:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.722465 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.722562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.722583 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.722640 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.722658 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:53Z","lastTransitionTime":"2025-10-01T15:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.807456 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:53 crc kubenswrapper[4726]: E1001 15:59:53.807652 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.825000 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.825073 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.825087 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.825107 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.825119 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:53Z","lastTransitionTime":"2025-10-01T15:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.927755 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.927789 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.927801 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.927818 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:53 crc kubenswrapper[4726]: I1001 15:59:53.927831 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:53Z","lastTransitionTime":"2025-10-01T15:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.030584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.030638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.030653 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.030675 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.030691 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:54Z","lastTransitionTime":"2025-10-01T15:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.134113 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.134167 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.134175 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.134192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.134202 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:54Z","lastTransitionTime":"2025-10-01T15:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.236875 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.236937 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.236954 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.236983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.237000 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:54Z","lastTransitionTime":"2025-10-01T15:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.339581 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.339624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.339632 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.339645 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.339656 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:54Z","lastTransitionTime":"2025-10-01T15:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.442558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.442615 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.442626 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.442644 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.442656 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:54Z","lastTransitionTime":"2025-10-01T15:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.545782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.545836 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.545848 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.545866 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.545880 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:54Z","lastTransitionTime":"2025-10-01T15:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.648417 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.648478 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.648496 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.648521 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.648538 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:54Z","lastTransitionTime":"2025-10-01T15:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.751599 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.751657 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.751676 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.751700 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.751720 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:54Z","lastTransitionTime":"2025-10-01T15:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.807513 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.807554 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.807528 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:54 crc kubenswrapper[4726]: E1001 15:59:54.807733 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 15:59:54 crc kubenswrapper[4726]: E1001 15:59:54.807879 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:54 crc kubenswrapper[4726]: E1001 15:59:54.807991 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.853652 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.853686 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.853695 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.853708 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.853724 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:54Z","lastTransitionTime":"2025-10-01T15:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.956283 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.956325 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.956335 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.956348 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:54 crc kubenswrapper[4726]: I1001 15:59:54.956359 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:54Z","lastTransitionTime":"2025-10-01T15:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.059304 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.059347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.059356 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.059372 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.059382 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:55Z","lastTransitionTime":"2025-10-01T15:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.161490 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.161559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.161577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.161601 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.161618 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:55Z","lastTransitionTime":"2025-10-01T15:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.264559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.264619 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.264638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.264662 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.264682 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:55Z","lastTransitionTime":"2025-10-01T15:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.367027 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.367138 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.367149 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.367165 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.367176 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:55Z","lastTransitionTime":"2025-10-01T15:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.469671 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.469712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.469723 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.469739 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.469750 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:55Z","lastTransitionTime":"2025-10-01T15:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.571844 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.572237 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.572385 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.573098 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.573259 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:55Z","lastTransitionTime":"2025-10-01T15:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.676257 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.676327 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.676349 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.676376 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.676397 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:55Z","lastTransitionTime":"2025-10-01T15:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.779734 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.779779 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.779792 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.779810 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.779824 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:55Z","lastTransitionTime":"2025-10-01T15:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.807667 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:55 crc kubenswrapper[4726]: E1001 15:59:55.807900 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.881995 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.882036 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.882065 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.882079 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.882089 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:55Z","lastTransitionTime":"2025-10-01T15:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.984626 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.984689 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.984706 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.984728 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:55 crc kubenswrapper[4726]: I1001 15:59:55.984745 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:55Z","lastTransitionTime":"2025-10-01T15:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.086815 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.086858 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.086870 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.086887 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.086899 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:56Z","lastTransitionTime":"2025-10-01T15:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.189023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.189109 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.189122 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.189140 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.189152 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:56Z","lastTransitionTime":"2025-10-01T15:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.292668 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.292704 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.292714 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.292727 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.292735 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:56Z","lastTransitionTime":"2025-10-01T15:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.395243 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.395514 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.395614 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.395704 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.395794 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:56Z","lastTransitionTime":"2025-10-01T15:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.498091 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.498161 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.498194 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.498215 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.498226 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:56Z","lastTransitionTime":"2025-10-01T15:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.601098 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.601438 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.601547 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.601647 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.601739 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:56Z","lastTransitionTime":"2025-10-01T15:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.704524 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.704813 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.704900 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.705008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.705125 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:56Z","lastTransitionTime":"2025-10-01T15:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.807465 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.807465 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:56 crc kubenswrapper[4726]: E1001 15:59:56.807737 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.807497 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:56 crc kubenswrapper[4726]: E1001 15:59:56.807826 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 15:59:56 crc kubenswrapper[4726]: E1001 15:59:56.807869 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.808214 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.808261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.808274 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.808292 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.808306 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:56Z","lastTransitionTime":"2025-10-01T15:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.910336 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.910376 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.910386 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.910400 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:56 crc kubenswrapper[4726]: I1001 15:59:56.910410 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:56Z","lastTransitionTime":"2025-10-01T15:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.012365 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.012419 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.012428 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.012441 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.012467 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:57Z","lastTransitionTime":"2025-10-01T15:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.115625 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.115720 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.115740 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.115763 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.115781 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:57Z","lastTransitionTime":"2025-10-01T15:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.217686 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.217762 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.217775 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.217790 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.217800 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:57Z","lastTransitionTime":"2025-10-01T15:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.319824 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.319859 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.319868 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.319880 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.319888 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:57Z","lastTransitionTime":"2025-10-01T15:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.422769 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.422799 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.422808 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.422821 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.422830 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:57Z","lastTransitionTime":"2025-10-01T15:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.525604 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.525681 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.525702 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.525727 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.525745 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:57Z","lastTransitionTime":"2025-10-01T15:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.627902 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.627948 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.627962 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.627977 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.627988 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:57Z","lastTransitionTime":"2025-10-01T15:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.730772 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.730840 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.730854 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.730870 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.730882 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:57Z","lastTransitionTime":"2025-10-01T15:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.807542 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:57 crc kubenswrapper[4726]: E1001 15:59:57.807699 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.832588 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:57Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.833309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.833352 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.833374 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.833406 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.833428 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:57Z","lastTransitionTime":"2025-10-01T15:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.851822 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:57Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.869040 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:57Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.886890 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:57Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.902930 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:57Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.925710 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"message\\\":\\\"169],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:59:43.163371 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:59:43.163377 6156 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1001 15:59:43.163391 6156 services_controller.go:443] Built service openshift-kube-scheduler/scheduler LB cluster-wide conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:57Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.935834 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.935877 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.935890 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.935908 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.935921 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:57Z","lastTransitionTime":"2025-10-01T15:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.941655 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:57Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.962702 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:57Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.976372 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:57Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:57 crc kubenswrapper[4726]: I1001 15:59:57.990395 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:57Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.002836 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:58Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.023390 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:58Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.036279 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:58Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.039654 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.039703 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.039715 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.039730 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.039741 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:58Z","lastTransitionTime":"2025-10-01T15:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.045797 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:58Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.055903 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:58Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.065208 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:58Z is after 2025-08-24T17:21:41Z" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.141846 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.141887 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.141897 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.141910 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.141919 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:58Z","lastTransitionTime":"2025-10-01T15:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.244109 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.244142 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.244152 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.244164 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.244173 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:58Z","lastTransitionTime":"2025-10-01T15:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.346637 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.346701 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.346719 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.346744 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.346761 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:58Z","lastTransitionTime":"2025-10-01T15:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.449543 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.449570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.449578 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.449590 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.449600 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:58Z","lastTransitionTime":"2025-10-01T15:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.552422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.552467 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.552476 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.552492 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.552501 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:58Z","lastTransitionTime":"2025-10-01T15:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.654814 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.654878 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.654891 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.654911 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.654925 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:58Z","lastTransitionTime":"2025-10-01T15:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.758188 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.758266 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.758286 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.758309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.758330 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:58Z","lastTransitionTime":"2025-10-01T15:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.808212 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.808291 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.808250 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:59:58 crc kubenswrapper[4726]: E1001 15:59:58.808525 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:59:58 crc kubenswrapper[4726]: E1001 15:59:58.808726 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:59:58 crc kubenswrapper[4726]: E1001 15:59:58.808866 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.861372 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.861453 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.861475 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.861501 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.861521 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:58Z","lastTransitionTime":"2025-10-01T15:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.964453 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.964498 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.964509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.964525 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:58 crc kubenswrapper[4726]: I1001 15:59:58.964538 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:58Z","lastTransitionTime":"2025-10-01T15:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.067389 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.067429 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.067437 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.067449 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.067458 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:59Z","lastTransitionTime":"2025-10-01T15:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.169493 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.169537 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.169546 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.169559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.169569 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:59Z","lastTransitionTime":"2025-10-01T15:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.272482 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.272528 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.272537 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.272550 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.272560 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:59Z","lastTransitionTime":"2025-10-01T15:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.375459 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.375534 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.375558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.375591 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.375615 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:59Z","lastTransitionTime":"2025-10-01T15:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.478687 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.478739 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.478751 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.478768 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.478781 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:59Z","lastTransitionTime":"2025-10-01T15:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.581799 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.581847 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.581858 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.581913 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.581927 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:59Z","lastTransitionTime":"2025-10-01T15:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.684679 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.684716 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.684724 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.684736 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.684745 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:59Z","lastTransitionTime":"2025-10-01T15:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.787504 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.787577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.787595 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.787624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.787641 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:59Z","lastTransitionTime":"2025-10-01T15:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.807070 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:59:59 crc kubenswrapper[4726]: E1001 15:59:59.807187 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.890311 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.890347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.890360 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.890375 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.890387 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:59Z","lastTransitionTime":"2025-10-01T15:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.993281 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.993324 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.993333 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.993346 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:59:59 crc kubenswrapper[4726]: I1001 15:59:59.993355 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:59:59Z","lastTransitionTime":"2025-10-01T15:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.096932 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.096970 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.096984 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.097001 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.097014 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:00Z","lastTransitionTime":"2025-10-01T16:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.199822 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.199863 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.199877 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.199897 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.199912 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:00Z","lastTransitionTime":"2025-10-01T16:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.302735 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.302779 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.302788 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.302801 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.302809 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:00Z","lastTransitionTime":"2025-10-01T16:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.404893 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.404943 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.404954 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.404973 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.404988 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:00Z","lastTransitionTime":"2025-10-01T16:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.507677 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.507721 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.507734 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.507773 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.507796 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:00Z","lastTransitionTime":"2025-10-01T16:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.611662 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.611715 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.611728 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.611745 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.611761 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:00Z","lastTransitionTime":"2025-10-01T16:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.717819 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.718762 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.718816 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.718835 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.718844 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:00Z","lastTransitionTime":"2025-10-01T16:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.807207 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.807292 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.807364 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:00 crc kubenswrapper[4726]: E1001 16:00:00.808030 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:00 crc kubenswrapper[4726]: E1001 16:00:00.808168 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:00 crc kubenswrapper[4726]: E1001 16:00:00.808211 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.822367 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.822707 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.822906 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.823471 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.823941 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:00Z","lastTransitionTime":"2025-10-01T16:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.927663 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.927752 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.927770 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.927794 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:00 crc kubenswrapper[4726]: I1001 16:00:00.927811 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:00Z","lastTransitionTime":"2025-10-01T16:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.030677 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.030726 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.030736 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.030750 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.030758 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:01Z","lastTransitionTime":"2025-10-01T16:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.133236 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.133306 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.133331 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.133360 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.133384 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:01Z","lastTransitionTime":"2025-10-01T16:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.235951 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.235992 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.236004 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.236019 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.236028 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:01Z","lastTransitionTime":"2025-10-01T16:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.326933 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs\") pod \"network-metrics-daemon-4pzd8\" (UID: \"2a22e7a8-726a-4538-a27d-44c58f89f0b1\") " pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.327097 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.327142 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs podName:2a22e7a8-726a-4538-a27d-44c58f89f0b1 nodeName:}" failed. No retries permitted until 2025-10-01 16:00:17.327127907 +0000 UTC m=+70.228680484 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs") pod "network-metrics-daemon-4pzd8" (UID: "2a22e7a8-726a-4538-a27d-44c58f89f0b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.338577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.338622 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.338637 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.338657 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.338672 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:01Z","lastTransitionTime":"2025-10-01T16:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.441366 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.441413 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.441427 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.441444 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.441460 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:01Z","lastTransitionTime":"2025-10-01T16:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.543904 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.543957 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.543969 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.543982 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.543991 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:01Z","lastTransitionTime":"2025-10-01T16:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.628793 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.628947 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.628997 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:00:33.628973009 +0000 UTC m=+86.530525586 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.629135 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.629162 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.629178 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.629281 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 16:00:33.629263066 +0000 UTC m=+86.530815653 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.646426 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.646480 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.646495 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.646515 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.646529 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:01Z","lastTransitionTime":"2025-10-01T16:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.683731 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.683800 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.683818 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.683843 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.683860 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:01Z","lastTransitionTime":"2025-10-01T16:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.704171 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:01Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.710025 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.710128 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.710147 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.710173 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.710191 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:01Z","lastTransitionTime":"2025-10-01T16:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.730620 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.730720 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.730808 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.730928 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.730809 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.730950 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 16:00:33.73092323 +0000 UTC m=+86.632475927 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.731015 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.731153 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 16:00:33.731126465 +0000 UTC m=+86.632679102 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.731164 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.731191 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.731284 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 16:00:33.731267519 +0000 UTC m=+86.632820166 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.732390 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:01Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.738099 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.738184 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.738228 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.738252 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.738269 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:01Z","lastTransitionTime":"2025-10-01T16:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.757586 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:01Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.761868 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.761924 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.761937 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.761960 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.761973 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:01Z","lastTransitionTime":"2025-10-01T16:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.775642 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:01Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.780125 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.780370 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.780436 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.780511 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.780569 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:01Z","lastTransitionTime":"2025-10-01T16:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.797520 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:01Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.797682 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.799322 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.799371 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.799383 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.799401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.799418 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:01Z","lastTransitionTime":"2025-10-01T16:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.807810 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:01 crc kubenswrapper[4726]: E1001 16:00:01.807931 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.808664 4726 scope.go:117] "RemoveContainer" containerID="7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.902423 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.902753 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.902765 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.902784 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:01 crc kubenswrapper[4726]: I1001 16:00:01.902799 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:01Z","lastTransitionTime":"2025-10-01T16:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.007185 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.007219 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.007228 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.007242 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.007267 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:02Z","lastTransitionTime":"2025-10-01T16:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.110505 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.110545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.110556 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.110571 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.110652 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:02Z","lastTransitionTime":"2025-10-01T16:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.175657 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/1.log" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.179641 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerStarted","Data":"63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9"} Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.180437 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.200156 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.212688 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.213713 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.213761 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.213770 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.213785 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.213794 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:02Z","lastTransitionTime":"2025-10-01T16:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.232092 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.245085 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.257634 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.268565 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.282659 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.299224 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.315976 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.316017 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.316030 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.316058 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.316072 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:02Z","lastTransitionTime":"2025-10-01T16:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.318589 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.330792 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.348522 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"message\\\":\\\"169],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:59:43.163371 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:59:43.163377 6156 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1001 15:59:43.163391 6156 services_controller.go:443] Built service openshift-kube-scheduler/scheduler LB cluster-wide conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T16:00:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.361825 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.371318 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.383302 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.395196 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.412517 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.418256 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.418294 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.418305 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.418324 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.418335 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:02Z","lastTransitionTime":"2025-10-01T16:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.520078 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.520116 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.520125 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.520140 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.520152 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:02Z","lastTransitionTime":"2025-10-01T16:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.623225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.623252 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.623261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.623274 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.623284 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:02Z","lastTransitionTime":"2025-10-01T16:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.726198 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.726258 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.726281 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.726309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.726326 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:02Z","lastTransitionTime":"2025-10-01T16:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.807284 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:02 crc kubenswrapper[4726]: E1001 16:00:02.807412 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.807470 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:02 crc kubenswrapper[4726]: E1001 16:00:02.807521 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.807565 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:02 crc kubenswrapper[4726]: E1001 16:00:02.807608 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.828661 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.828696 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.828708 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.828725 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.828736 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:02Z","lastTransitionTime":"2025-10-01T16:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.932142 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.932236 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.932262 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.932962 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:02 crc kubenswrapper[4726]: I1001 16:00:02.933427 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:02Z","lastTransitionTime":"2025-10-01T16:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.036995 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.037399 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.037412 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.037431 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.037443 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:03Z","lastTransitionTime":"2025-10-01T16:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.083068 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.099033 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.106085 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.119923 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.132393 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.141080 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.141415 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.141426 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.141439 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.141450 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:03Z","lastTransitionTime":"2025-10-01T16:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.142589 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.154029 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.165343 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.177037 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.185219 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/2.log" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.186033 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/1.log" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.188348 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.190073 4726 generic.go:334] "Generic (PLEG): container finished" podID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerID="63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9" exitCode=1 Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.190265 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerDied","Data":"63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9"} Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.190342 4726 scope.go:117] "RemoveContainer" containerID="7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.190767 4726 scope.go:117] "RemoveContainer" containerID="63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9" Oct 01 16:00:03 crc kubenswrapper[4726]: E1001 16:00:03.190952 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.200342 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.211036 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.228748 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"message\\\":\\\"169],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:59:43.163371 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:59:43.163377 6156 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1001 15:59:43.163391 6156 services_controller.go:443] Built service openshift-kube-scheduler/scheduler LB cluster-wide conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T16:00:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.242551 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.243216 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.243243 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.243253 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.243269 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.243279 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:03Z","lastTransitionTime":"2025-10-01T16:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.254288 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.263875 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.273756 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.282483 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.292065 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.301694 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.316432 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.328409 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.338601 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.347284 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.347312 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.347320 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.347332 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.347341 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:03Z","lastTransitionTime":"2025-10-01T16:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.350831 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.361311 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.369705 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.379488 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.391357 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.405341 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.415341 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.436915 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cb93c3d68326c5ffe6c8944faa6133b96394c3f1d02def07bb4b5d62eed48cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"message\\\":\\\"169],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:59:43.163371 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:59:43Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:59:43.163377 6156 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1001 15:59:43.163391 6156 services_controller.go:443] Built service openshift-kube-scheduler/scheduler LB cluster-wide conf\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:02Z\\\",\\\"message\\\":\\\"oller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z]\\\\nI1001 16:00:02.627548 6407 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager-operator/metrics for network=default are: map[]\\\\nI1001 16:00:02.627560 6407 services_controller.go:443] Built service openshift-kube-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.219\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1001 16:00:02.627572 6407 services_controller.go:444] Built service opens\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T16:00:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.450356 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.450395 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.450404 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.450425 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.450435 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:03Z","lastTransitionTime":"2025-10-01T16:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.455350 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.468228 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12eb575b-72e7-4bb2-b9de-ce924d47db1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9fe86b3d6a1acd9512663da493ea130d4112b79d2ea24e5d94eeb76e81d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f94a8081041e5968c831dce9988a771386310ae721226c4f58baa6a2c1af47b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565eeb4bdc62a6f23f5a056700c8677e7de8bac37bd7cc766d4b8637c7fe3f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.483884 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.497158 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:03Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.553391 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.553438 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.553452 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.553473 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.553486 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:03Z","lastTransitionTime":"2025-10-01T16:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.656666 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.656767 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.656791 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.657534 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.658131 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:03Z","lastTransitionTime":"2025-10-01T16:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.761194 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.761819 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.761842 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.761859 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.761870 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:03Z","lastTransitionTime":"2025-10-01T16:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.807889 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:03 crc kubenswrapper[4726]: E1001 16:00:03.808028 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.863317 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.863348 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.863356 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.863369 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.863378 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:03Z","lastTransitionTime":"2025-10-01T16:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.968625 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.968673 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.968684 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.968703 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:03 crc kubenswrapper[4726]: I1001 16:00:03.968715 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:03Z","lastTransitionTime":"2025-10-01T16:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.071390 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.071431 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.071439 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.071454 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.071466 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:04Z","lastTransitionTime":"2025-10-01T16:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.173601 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.173643 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.173655 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.173673 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.173684 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:04Z","lastTransitionTime":"2025-10-01T16:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.194875 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/2.log" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.198393 4726 scope.go:117] "RemoveContainer" containerID="63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9" Oct 01 16:00:04 crc kubenswrapper[4726]: E1001 16:00:04.198566 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.211766 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12eb575b-72e7-4bb2-b9de-ce924d47db1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9fe86b3d6a1acd9512663da493ea130d4112b79d2ea24e5d94eeb76e81d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f94a8081041e5968c831dce9988a771386310ae721226c4f58baa6a2c1af47b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565eeb4bdc62a6f23f5a056700c8677e7de8bac37bd7cc766d4b8637c7fe3f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.226610 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.243008 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.255213 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.268619 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.275745 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.275801 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.275823 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.275846 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.275861 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:04Z","lastTransitionTime":"2025-10-01T16:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.289585 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:02Z\\\",\\\"message\\\":\\\"oller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z]\\\\nI1001 16:00:02.627548 6407 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager-operator/metrics for network=default are: map[]\\\\nI1001 16:00:02.627560 6407 services_controller.go:443] Built service openshift-kube-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.219\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1001 16:00:02.627572 6407 services_controller.go:444] Built service opens\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T16:00:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.303192 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.316261 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.326422 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.338513 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.349033 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.362896 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.377197 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.378517 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.378545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.378555 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.378584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.378594 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:04Z","lastTransitionTime":"2025-10-01T16:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.388967 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.401526 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.411233 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.420788 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.480851 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.480890 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.480899 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.480913 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.480923 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:04Z","lastTransitionTime":"2025-10-01T16:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.583329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.583366 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.583376 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.583390 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.583398 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:04Z","lastTransitionTime":"2025-10-01T16:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.687834 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.687893 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.687924 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.687946 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.687961 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:04Z","lastTransitionTime":"2025-10-01T16:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.693620 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.712895 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.729820 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.743586 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.766276 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:02Z\\\",\\\"message\\\":\\\"oller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z]\\\\nI1001 16:00:02.627548 6407 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager-operator/metrics for network=default are: map[]\\\\nI1001 16:00:02.627560 6407 services_controller.go:443] Built service openshift-kube-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.219\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1001 16:00:02.627572 6407 services_controller.go:444] Built service opens\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T16:00:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.781121 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.791245 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.791323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.791351 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.791378 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.791401 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:04Z","lastTransitionTime":"2025-10-01T16:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.798260 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12eb575b-72e7-4bb2-b9de-ce924d47db1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9fe86b3d6a1acd9512663da493ea130d4112b79d2ea24e5d94eeb76e81d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f94a8081041e5968c831dce9988a771386310ae721226c4f58baa6a2c1af47b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565eeb4bdc62a6f23f5a056700c8677e7de8bac37bd7cc766d4b8637c7fe3f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.807771 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.807795 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.807859 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:04 crc kubenswrapper[4726]: E1001 16:00:04.807900 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:04 crc kubenswrapper[4726]: E1001 16:00:04.808006 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:04 crc kubenswrapper[4726]: E1001 16:00:04.808167 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.814915 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.828829 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.845360 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.858291 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.868586 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.880564 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.894698 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.894767 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.894778 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.894810 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.894823 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:04Z","lastTransitionTime":"2025-10-01T16:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.895965 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.909710 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.922320 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.932804 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.946382 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:04Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.998023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.998119 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.998138 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.998165 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:04 crc kubenswrapper[4726]: I1001 16:00:04.998182 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:04Z","lastTransitionTime":"2025-10-01T16:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.101017 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.101091 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.101107 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.101124 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.101134 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:05Z","lastTransitionTime":"2025-10-01T16:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.203606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.203650 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.203683 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.203702 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.203716 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:05Z","lastTransitionTime":"2025-10-01T16:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.307351 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.307408 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.307428 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.307451 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.307469 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:05Z","lastTransitionTime":"2025-10-01T16:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.410622 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.410670 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.410682 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.410709 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.410721 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:05Z","lastTransitionTime":"2025-10-01T16:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.513702 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.513764 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.513779 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.513800 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.513817 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:05Z","lastTransitionTime":"2025-10-01T16:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.616395 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.616463 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.616478 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.616499 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.616510 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:05Z","lastTransitionTime":"2025-10-01T16:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.718499 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.718596 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.718606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.718628 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.718641 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:05Z","lastTransitionTime":"2025-10-01T16:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.807092 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:05 crc kubenswrapper[4726]: E1001 16:00:05.807236 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.821118 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.821185 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.821200 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.821223 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.821242 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:05Z","lastTransitionTime":"2025-10-01T16:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.924158 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.924251 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.924276 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.924296 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:05 crc kubenswrapper[4726]: I1001 16:00:05.924308 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:05Z","lastTransitionTime":"2025-10-01T16:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.027142 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.027179 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.027189 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.027205 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.027217 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:06Z","lastTransitionTime":"2025-10-01T16:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.129563 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.129618 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.129638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.129655 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.129667 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:06Z","lastTransitionTime":"2025-10-01T16:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.231695 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.231826 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.231841 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.231856 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.231866 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:06Z","lastTransitionTime":"2025-10-01T16:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.334151 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.334192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.334210 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.334229 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.334243 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:06Z","lastTransitionTime":"2025-10-01T16:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.436607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.436646 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.436654 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.436666 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.436674 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:06Z","lastTransitionTime":"2025-10-01T16:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.540098 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.540143 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.540159 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.540176 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.540188 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:06Z","lastTransitionTime":"2025-10-01T16:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.643303 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.643375 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.643398 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.643426 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.643444 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:06Z","lastTransitionTime":"2025-10-01T16:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.746702 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.746756 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.746765 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.746779 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.746791 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:06Z","lastTransitionTime":"2025-10-01T16:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.807133 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.807228 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.807149 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:06 crc kubenswrapper[4726]: E1001 16:00:06.807331 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:06 crc kubenswrapper[4726]: E1001 16:00:06.807440 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:06 crc kubenswrapper[4726]: E1001 16:00:06.807525 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.849197 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.849255 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.849267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.849284 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.849296 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:06Z","lastTransitionTime":"2025-10-01T16:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.953554 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.953593 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.953604 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.953620 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:06 crc kubenswrapper[4726]: I1001 16:00:06.953631 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:06Z","lastTransitionTime":"2025-10-01T16:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.056701 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.056781 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.056813 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.056846 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.056875 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:07Z","lastTransitionTime":"2025-10-01T16:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.159544 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.159580 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.159588 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.159602 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.159611 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:07Z","lastTransitionTime":"2025-10-01T16:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.262851 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.262951 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.262970 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.263031 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.263100 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:07Z","lastTransitionTime":"2025-10-01T16:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.368212 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.368311 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.368329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.368357 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.368386 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:07Z","lastTransitionTime":"2025-10-01T16:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.471267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.471326 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.471345 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.471368 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.471384 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:07Z","lastTransitionTime":"2025-10-01T16:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.574658 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.574739 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.574759 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.574812 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.574828 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:07Z","lastTransitionTime":"2025-10-01T16:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.678811 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.678890 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.678912 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.678942 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.678965 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:07Z","lastTransitionTime":"2025-10-01T16:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.782359 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.782430 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.782448 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.782473 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.782491 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:07Z","lastTransitionTime":"2025-10-01T16:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.807253 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:07 crc kubenswrapper[4726]: E1001 16:00:07.807417 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.821173 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:07Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.838165 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:07Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.852334 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:07Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.864458 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:07Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.874831 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:07Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.885303 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.885347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.885358 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.885374 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.885385 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:07Z","lastTransitionTime":"2025-10-01T16:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.901622 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:02Z\\\",\\\"message\\\":\\\"oller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z]\\\\nI1001 16:00:02.627548 6407 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager-operator/metrics for network=default are: map[]\\\\nI1001 16:00:02.627560 6407 services_controller.go:443] Built service openshift-kube-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.219\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1001 16:00:02.627572 6407 services_controller.go:444] Built service opens\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T16:00:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:07Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.919828 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:07Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.934666 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12eb575b-72e7-4bb2-b9de-ce924d47db1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9fe86b3d6a1acd9512663da493ea130d4112b79d2ea24e5d94eeb76e81d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f94a8081041e5968c831dce9988a771386310ae721226c4f58baa6a2c1af47b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565eeb4bdc62a6f23f5a056700c8677e7de8bac37bd7cc766d4b8637c7fe3f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:07Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.949000 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:07Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.959943 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:07Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.970958 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:07Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.983477 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:07Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.987614 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.987671 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.987684 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.987702 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.987713 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:07Z","lastTransitionTime":"2025-10-01T16:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:07 crc kubenswrapper[4726]: I1001 16:00:07.992710 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:07Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.000769 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:07Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.011724 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:08Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.021921 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:08Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.032465 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:08Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.089568 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.089643 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.089663 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.089686 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.089703 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:08Z","lastTransitionTime":"2025-10-01T16:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.192384 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.192530 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.192587 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.192618 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.192638 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:08Z","lastTransitionTime":"2025-10-01T16:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.294813 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.294854 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.294866 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.294881 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.294893 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:08Z","lastTransitionTime":"2025-10-01T16:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.397340 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.397400 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.397440 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.397474 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.397497 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:08Z","lastTransitionTime":"2025-10-01T16:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.499554 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.499610 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.499624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.499644 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.499659 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:08Z","lastTransitionTime":"2025-10-01T16:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.602735 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.602822 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.602839 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.602861 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.602876 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:08Z","lastTransitionTime":"2025-10-01T16:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.705303 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.705365 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.705376 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.705390 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.705401 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:08Z","lastTransitionTime":"2025-10-01T16:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.807114 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.807138 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:08 crc kubenswrapper[4726]: E1001 16:00:08.807225 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.807115 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:08 crc kubenswrapper[4726]: E1001 16:00:08.807345 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:08 crc kubenswrapper[4726]: E1001 16:00:08.807455 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.807516 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.807585 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.807621 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.807651 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.807677 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:08Z","lastTransitionTime":"2025-10-01T16:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.910820 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.910890 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.910913 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.910943 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:08 crc kubenswrapper[4726]: I1001 16:00:08.910966 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:08Z","lastTransitionTime":"2025-10-01T16:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.014146 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.014195 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.014208 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.014226 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.014240 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:09Z","lastTransitionTime":"2025-10-01T16:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.117782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.117824 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.117834 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.117850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.117860 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:09Z","lastTransitionTime":"2025-10-01T16:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.219901 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.219966 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.219981 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.220005 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.220021 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:09Z","lastTransitionTime":"2025-10-01T16:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.323791 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.323843 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.323855 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.323878 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.323890 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:09Z","lastTransitionTime":"2025-10-01T16:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.426357 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.426410 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.426422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.426442 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.426457 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:09Z","lastTransitionTime":"2025-10-01T16:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.529722 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.529797 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.529817 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.529872 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.529890 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:09Z","lastTransitionTime":"2025-10-01T16:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.632524 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.632588 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.632596 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.632611 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.632620 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:09Z","lastTransitionTime":"2025-10-01T16:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.737296 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.737386 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.737408 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.737435 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.737452 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:09Z","lastTransitionTime":"2025-10-01T16:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.808191 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:09 crc kubenswrapper[4726]: E1001 16:00:09.808413 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.845574 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.845642 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.845662 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.845691 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.845710 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:09Z","lastTransitionTime":"2025-10-01T16:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.949173 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.949264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.949283 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.949309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:09 crc kubenswrapper[4726]: I1001 16:00:09.949326 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:09Z","lastTransitionTime":"2025-10-01T16:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.052406 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.052464 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.052478 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.052502 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.052517 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:10Z","lastTransitionTime":"2025-10-01T16:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.156641 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.156711 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.156722 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.156739 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.156750 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:10Z","lastTransitionTime":"2025-10-01T16:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.259550 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.259597 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.259606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.259622 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.259633 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:10Z","lastTransitionTime":"2025-10-01T16:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.362589 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.362636 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.362649 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.362665 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.362678 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:10Z","lastTransitionTime":"2025-10-01T16:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.465520 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.465589 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.465601 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.465622 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.465655 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:10Z","lastTransitionTime":"2025-10-01T16:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.569846 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.569895 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.569905 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.569922 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.569936 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:10Z","lastTransitionTime":"2025-10-01T16:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.672920 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.673001 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.673025 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.673107 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.673134 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:10Z","lastTransitionTime":"2025-10-01T16:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.775864 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.775897 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.775906 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.775920 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.775929 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:10Z","lastTransitionTime":"2025-10-01T16:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.807885 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.807966 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.807901 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:10 crc kubenswrapper[4726]: E1001 16:00:10.808023 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:10 crc kubenswrapper[4726]: E1001 16:00:10.808098 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:10 crc kubenswrapper[4726]: E1001 16:00:10.808164 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.878483 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.878540 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.878549 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.878562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.878577 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:10Z","lastTransitionTime":"2025-10-01T16:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.981165 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.981230 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.981248 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.981274 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:10 crc kubenswrapper[4726]: I1001 16:00:10.981291 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:10Z","lastTransitionTime":"2025-10-01T16:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.083814 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.083887 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.083905 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.083931 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.083959 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:11Z","lastTransitionTime":"2025-10-01T16:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.187426 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.187507 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.187532 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.187560 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.187578 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:11Z","lastTransitionTime":"2025-10-01T16:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.290504 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.290537 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.290545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.290556 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.290565 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:11Z","lastTransitionTime":"2025-10-01T16:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.393154 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.393185 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.393193 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.393205 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.393214 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:11Z","lastTransitionTime":"2025-10-01T16:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.495898 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.496000 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.496022 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.496077 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.496102 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:11Z","lastTransitionTime":"2025-10-01T16:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.603110 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.603166 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.603184 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.603207 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.603223 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:11Z","lastTransitionTime":"2025-10-01T16:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.706088 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.706124 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.706135 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.706149 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.706161 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:11Z","lastTransitionTime":"2025-10-01T16:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.807336 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:11 crc kubenswrapper[4726]: E1001 16:00:11.807538 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.808997 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.809095 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.809118 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.809146 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.809168 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:11Z","lastTransitionTime":"2025-10-01T16:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.911619 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.911669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.911681 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.911700 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:11 crc kubenswrapper[4726]: I1001 16:00:11.911711 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:11Z","lastTransitionTime":"2025-10-01T16:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.013469 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.013511 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.013524 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.013544 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.013557 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.116234 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.116281 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.116293 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.116309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.116321 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.137031 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.137099 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.137114 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.137130 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.137141 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: E1001 16:00:12.174932 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:12Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.181942 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.182023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.182040 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.182094 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.182113 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: E1001 16:00:12.205916 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:12Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.210784 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.210837 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.210848 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.210864 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.210875 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: E1001 16:00:12.227086 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:12Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.231279 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.231318 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.231329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.231345 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.231360 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: E1001 16:00:12.245155 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:12Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.248719 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.248758 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.248766 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.248781 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.248793 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: E1001 16:00:12.263355 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:12Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:12 crc kubenswrapper[4726]: E1001 16:00:12.263473 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.264860 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.264908 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.264920 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.264937 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.264950 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.368445 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.368484 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.368495 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.368509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.368518 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.471264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.471302 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.471311 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.471325 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.471338 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.574238 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.574269 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.574276 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.574305 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.574314 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.676317 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.676354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.676363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.676377 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.676386 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.779764 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.779819 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.779830 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.779846 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.779860 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.807931 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.808003 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:12 crc kubenswrapper[4726]: E1001 16:00:12.808173 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.808198 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:12 crc kubenswrapper[4726]: E1001 16:00:12.808290 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:12 crc kubenswrapper[4726]: E1001 16:00:12.808462 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.882110 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.882230 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.882311 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.882384 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.882411 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.985464 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.985509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.985520 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.985535 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:12 crc kubenswrapper[4726]: I1001 16:00:12.985547 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:12Z","lastTransitionTime":"2025-10-01T16:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.087485 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.087531 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.087544 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.087561 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.087573 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:13Z","lastTransitionTime":"2025-10-01T16:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.190169 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.190217 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.190230 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.190247 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.190259 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:13Z","lastTransitionTime":"2025-10-01T16:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.292792 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.293141 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.293154 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.293171 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.293183 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:13Z","lastTransitionTime":"2025-10-01T16:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.395479 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.395514 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.395525 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.395540 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.395552 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:13Z","lastTransitionTime":"2025-10-01T16:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.498376 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.498434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.498444 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.498464 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.498477 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:13Z","lastTransitionTime":"2025-10-01T16:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.601164 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.601226 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.601238 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.601259 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.601272 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:13Z","lastTransitionTime":"2025-10-01T16:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.703445 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.703482 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.703491 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.703505 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.703515 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:13Z","lastTransitionTime":"2025-10-01T16:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.806085 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.806129 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.806141 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.806160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.806172 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:13Z","lastTransitionTime":"2025-10-01T16:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.807554 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:13 crc kubenswrapper[4726]: E1001 16:00:13.807675 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.908911 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.908954 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.908966 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.908983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:13 crc kubenswrapper[4726]: I1001 16:00:13.908994 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:13Z","lastTransitionTime":"2025-10-01T16:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.011858 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.011896 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.011907 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.011922 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.011933 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:14Z","lastTransitionTime":"2025-10-01T16:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.113840 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.113895 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.113906 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.113922 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.113933 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:14Z","lastTransitionTime":"2025-10-01T16:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.216192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.216224 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.216234 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.216248 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.216258 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:14Z","lastTransitionTime":"2025-10-01T16:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.318381 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.318433 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.318448 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.318466 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.318477 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:14Z","lastTransitionTime":"2025-10-01T16:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.420985 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.421023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.421032 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.421062 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.421072 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:14Z","lastTransitionTime":"2025-10-01T16:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.523851 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.523905 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.523924 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.523945 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.523963 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:14Z","lastTransitionTime":"2025-10-01T16:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.627243 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.627300 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.627310 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.627325 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.627336 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:14Z","lastTransitionTime":"2025-10-01T16:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.729328 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.729372 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.729385 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.729403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.729417 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:14Z","lastTransitionTime":"2025-10-01T16:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.807339 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.807382 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.807441 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:14 crc kubenswrapper[4726]: E1001 16:00:14.807466 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:14 crc kubenswrapper[4726]: E1001 16:00:14.807628 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:14 crc kubenswrapper[4726]: E1001 16:00:14.807662 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.832355 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.832394 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.832406 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.832422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.832432 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:14Z","lastTransitionTime":"2025-10-01T16:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.934911 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.934942 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.934953 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.934968 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:14 crc kubenswrapper[4726]: I1001 16:00:14.934979 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:14Z","lastTransitionTime":"2025-10-01T16:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.037590 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.037641 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.037654 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.037672 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.037682 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:15Z","lastTransitionTime":"2025-10-01T16:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.140341 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.140396 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.140413 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.140443 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.140460 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:15Z","lastTransitionTime":"2025-10-01T16:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.242512 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.242549 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.242558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.242574 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.242586 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:15Z","lastTransitionTime":"2025-10-01T16:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.345333 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.345599 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.345687 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.345763 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.345835 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:15Z","lastTransitionTime":"2025-10-01T16:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.448684 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.448735 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.448747 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.448766 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.448777 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:15Z","lastTransitionTime":"2025-10-01T16:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.550702 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.550748 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.550767 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.550791 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.550801 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:15Z","lastTransitionTime":"2025-10-01T16:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.653431 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.653457 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.653466 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.653478 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.653488 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:15Z","lastTransitionTime":"2025-10-01T16:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.756300 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.756337 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.756346 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.756361 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.756370 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:15Z","lastTransitionTime":"2025-10-01T16:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.807499 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:15 crc kubenswrapper[4726]: E1001 16:00:15.807631 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.858545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.858584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.858593 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.858607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.858618 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:15Z","lastTransitionTime":"2025-10-01T16:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.961445 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.961507 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.961522 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.961545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:15 crc kubenswrapper[4726]: I1001 16:00:15.961560 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:15Z","lastTransitionTime":"2025-10-01T16:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.063479 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.063540 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.063551 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.063567 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.063577 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:16Z","lastTransitionTime":"2025-10-01T16:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.165521 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.165566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.165576 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.165591 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.165602 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:16Z","lastTransitionTime":"2025-10-01T16:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.267570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.267625 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.267635 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.267649 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.267659 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:16Z","lastTransitionTime":"2025-10-01T16:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.369863 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.369921 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.369931 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.369961 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.369976 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:16Z","lastTransitionTime":"2025-10-01T16:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.472521 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.472570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.472579 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.472596 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.472607 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:16Z","lastTransitionTime":"2025-10-01T16:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.576021 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.576082 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.576092 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.576109 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.576122 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:16Z","lastTransitionTime":"2025-10-01T16:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.678647 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.678699 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.678711 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.678727 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.678736 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:16Z","lastTransitionTime":"2025-10-01T16:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.782227 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.782272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.782287 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.782306 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.782318 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:16Z","lastTransitionTime":"2025-10-01T16:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.807531 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.807546 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:16 crc kubenswrapper[4726]: E1001 16:00:16.807653 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.807721 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:16 crc kubenswrapper[4726]: E1001 16:00:16.807892 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:16 crc kubenswrapper[4726]: E1001 16:00:16.807989 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.885768 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.885807 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.885820 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.885839 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.885850 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:16Z","lastTransitionTime":"2025-10-01T16:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.988389 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.988423 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.988434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.988447 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:16 crc kubenswrapper[4726]: I1001 16:00:16.988457 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:16Z","lastTransitionTime":"2025-10-01T16:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.090420 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.090466 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.090477 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.090494 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.090505 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:17Z","lastTransitionTime":"2025-10-01T16:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.192706 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.192755 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.192768 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.192783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.192795 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:17Z","lastTransitionTime":"2025-10-01T16:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.295214 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.295280 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.295300 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.295327 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.295348 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:17Z","lastTransitionTime":"2025-10-01T16:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.392140 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs\") pod \"network-metrics-daemon-4pzd8\" (UID: \"2a22e7a8-726a-4538-a27d-44c58f89f0b1\") " pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:17 crc kubenswrapper[4726]: E1001 16:00:17.392253 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 16:00:17 crc kubenswrapper[4726]: E1001 16:00:17.392336 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs podName:2a22e7a8-726a-4538-a27d-44c58f89f0b1 nodeName:}" failed. No retries permitted until 2025-10-01 16:00:49.39231814 +0000 UTC m=+102.293870717 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs") pod "network-metrics-daemon-4pzd8" (UID: "2a22e7a8-726a-4538-a27d-44c58f89f0b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.397538 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.397565 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.397576 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.397590 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.397599 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:17Z","lastTransitionTime":"2025-10-01T16:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.500174 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.500253 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.500267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.500286 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.500298 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:17Z","lastTransitionTime":"2025-10-01T16:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.602507 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.602572 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.602584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.602618 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.602634 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:17Z","lastTransitionTime":"2025-10-01T16:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.704586 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.704624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.704633 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.704646 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.704655 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:17Z","lastTransitionTime":"2025-10-01T16:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.806758 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.806790 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.806801 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.806815 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.806825 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:17Z","lastTransitionTime":"2025-10-01T16:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.807047 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:17 crc kubenswrapper[4726]: E1001 16:00:17.807287 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.821068 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:17Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.835312 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:17Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.847646 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:17Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.864154 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:02Z\\\",\\\"message\\\":\\\"oller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z]\\\\nI1001 16:00:02.627548 6407 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager-operator/metrics for network=default are: map[]\\\\nI1001 16:00:02.627560 6407 services_controller.go:443] Built service openshift-kube-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.219\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1001 16:00:02.627572 6407 services_controller.go:444] Built service opens\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T16:00:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:17Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.881651 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:17Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.894402 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12eb575b-72e7-4bb2-b9de-ce924d47db1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9fe86b3d6a1acd9512663da493ea130d4112b79d2ea24e5d94eeb76e81d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f94a8081041e5968c831dce9988a771386310ae721226c4f58baa6a2c1af47b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565eeb4bdc62a6f23f5a056700c8677e7de8bac37bd7cc766d4b8637c7fe3f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:17Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.907626 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:17Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.912554 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.912598 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.912612 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.912629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.912647 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:17Z","lastTransitionTime":"2025-10-01T16:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.923201 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:17Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.934815 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:17Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.949643 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:17Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.961104 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:17Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.972069 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:17Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.983708 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:17Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:17 crc kubenswrapper[4726]: I1001 16:00:17.994197 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:17Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.005193 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:18Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.013447 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:18Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.014742 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.014778 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.014791 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.014808 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.014819 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:18Z","lastTransitionTime":"2025-10-01T16:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.023890 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:18Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.116699 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.116747 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.116757 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.116775 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.116786 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:18Z","lastTransitionTime":"2025-10-01T16:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.219442 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.219488 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.219501 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.219520 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.219533 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:18Z","lastTransitionTime":"2025-10-01T16:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.321795 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.321862 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.321880 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.321905 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.321923 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:18Z","lastTransitionTime":"2025-10-01T16:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.424267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.424316 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.424330 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.424348 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.424359 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:18Z","lastTransitionTime":"2025-10-01T16:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.526238 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.526298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.526316 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.526339 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.526356 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:18Z","lastTransitionTime":"2025-10-01T16:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.628527 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.628596 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.628623 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.628650 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.628671 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:18Z","lastTransitionTime":"2025-10-01T16:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.731909 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.731958 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.731975 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.731997 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.732013 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:18Z","lastTransitionTime":"2025-10-01T16:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.806911 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.807288 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.807322 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:18 crc kubenswrapper[4726]: E1001 16:00:18.808107 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:18 crc kubenswrapper[4726]: E1001 16:00:18.807725 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:18 crc kubenswrapper[4726]: E1001 16:00:18.807593 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.820592 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.834782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.835043 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.835199 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.835331 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.835460 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:18Z","lastTransitionTime":"2025-10-01T16:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.938101 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.938403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.938565 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.938736 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:18 crc kubenswrapper[4726]: I1001 16:00:18.938905 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:18Z","lastTransitionTime":"2025-10-01T16:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.042474 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.042840 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.043031 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.043255 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.043402 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:19Z","lastTransitionTime":"2025-10-01T16:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.145457 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.145494 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.145523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.145536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.145546 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:19Z","lastTransitionTime":"2025-10-01T16:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.246982 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.247022 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.247034 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.247064 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.247073 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:19Z","lastTransitionTime":"2025-10-01T16:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.247392 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h8cql_e9257aec-e319-4261-ae4c-dec88468b680/kube-multus/0.log" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.247429 4726 generic.go:334] "Generic (PLEG): container finished" podID="e9257aec-e319-4261-ae4c-dec88468b680" containerID="e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead" exitCode=1 Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.247514 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h8cql" event={"ID":"e9257aec-e319-4261-ae4c-dec88468b680","Type":"ContainerDied","Data":"e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead"} Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.250734 4726 scope.go:117] "RemoveContainer" containerID="e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.260236 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.273129 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.284071 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.298541 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.310060 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.323260 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.333750 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.343745 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.349439 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.349475 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.349483 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.349497 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.349506 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:19Z","lastTransitionTime":"2025-10-01T16:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.354526 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.364856 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.378095 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.390197 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12eb575b-72e7-4bb2-b9de-ce924d47db1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9fe86b3d6a1acd9512663da493ea130d4112b79d2ea24e5d94eeb76e81d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f94a8081041e5968c831dce9988a771386310ae721226c4f58baa6a2c1af47b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565eeb4bdc62a6f23f5a056700c8677e7de8bac37bd7cc766d4b8637c7fe3f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.402014 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.414005 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.424188 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:18Z\\\",\\\"message\\\":\\\"2025-10-01T15:59:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14\\\\n2025-10-01T15:59:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14 to /host/opt/cni/bin/\\\\n2025-10-01T15:59:33Z [verbose] multus-daemon started\\\\n2025-10-01T15:59:33Z [verbose] Readiness Indicator file check\\\\n2025-10-01T16:00:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.432685 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.447214 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:02Z\\\",\\\"message\\\":\\\"oller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z]\\\\nI1001 16:00:02.627548 6407 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager-operator/metrics for network=default are: map[]\\\\nI1001 16:00:02.627560 6407 services_controller.go:443] Built service openshift-kube-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.219\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1001 16:00:02.627572 6407 services_controller.go:444] Built service opens\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T16:00:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.451203 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.451235 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.451243 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.451258 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.451268 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:19Z","lastTransitionTime":"2025-10-01T16:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.455250 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3064c56-dd95-414e-83a1-8876f73866fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4179687cd0b6fa600a3efc7b5f5761a15fb6df23af0f1b0c94170bb106b3154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:19Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.554561 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.554604 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.554615 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.554629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.554639 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:19Z","lastTransitionTime":"2025-10-01T16:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.657181 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.657216 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.657224 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.657238 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.657248 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:19Z","lastTransitionTime":"2025-10-01T16:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.759868 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.759930 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.759947 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.759974 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.759992 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:19Z","lastTransitionTime":"2025-10-01T16:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.808407 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:19 crc kubenswrapper[4726]: E1001 16:00:19.808657 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.809647 4726 scope.go:117] "RemoveContainer" containerID="63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9" Oct 01 16:00:19 crc kubenswrapper[4726]: E1001 16:00:19.809883 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.862939 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.862978 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.862990 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.863007 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.863021 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:19Z","lastTransitionTime":"2025-10-01T16:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.965171 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.965208 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.965219 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.965234 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:19 crc kubenswrapper[4726]: I1001 16:00:19.965247 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:19Z","lastTransitionTime":"2025-10-01T16:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.067255 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.067288 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.067296 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.067307 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.067317 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:20Z","lastTransitionTime":"2025-10-01T16:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.169332 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.169366 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.169377 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.169395 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.169404 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:20Z","lastTransitionTime":"2025-10-01T16:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.252116 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h8cql_e9257aec-e319-4261-ae4c-dec88468b680/kube-multus/0.log" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.252168 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h8cql" event={"ID":"e9257aec-e319-4261-ae4c-dec88468b680","Type":"ContainerStarted","Data":"e127584cf73c7f9ea20da97dce5f50fb86b8f9fde6ae383172834d425bb1e033"} Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.266789 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.271465 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.271527 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.271552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.271579 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.271597 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:20Z","lastTransitionTime":"2025-10-01T16:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.281521 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.294227 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.307117 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.317121 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.331237 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.342622 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12eb575b-72e7-4bb2-b9de-ce924d47db1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9fe86b3d6a1acd9512663da493ea130d4112b79d2ea24e5d94eeb76e81d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f94a8081041e5968c831dce9988a771386310ae721226c4f58baa6a2c1af47b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565eeb4bdc62a6f23f5a056700c8677e7de8bac37bd7cc766d4b8637c7fe3f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.354379 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.367741 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.374135 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.374197 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.374209 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.374247 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.374259 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:20Z","lastTransitionTime":"2025-10-01T16:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.380811 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e127584cf73c7f9ea20da97dce5f50fb86b8f9fde6ae383172834d425bb1e033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:18Z\\\",\\\"message\\\":\\\"2025-10-01T15:59:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14\\\\n2025-10-01T15:59:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14 to /host/opt/cni/bin/\\\\n2025-10-01T15:59:33Z [verbose] multus-daemon started\\\\n2025-10-01T15:59:33Z [verbose] Readiness Indicator file check\\\\n2025-10-01T16:00:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T16:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.392490 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.410327 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:02Z\\\",\\\"message\\\":\\\"oller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z]\\\\nI1001 16:00:02.627548 6407 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager-operator/metrics for network=default are: map[]\\\\nI1001 16:00:02.627560 6407 services_controller.go:443] Built service openshift-kube-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.219\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1001 16:00:02.627572 6407 services_controller.go:444] Built service opens\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T16:00:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.420371 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3064c56-dd95-414e-83a1-8876f73866fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4179687cd0b6fa600a3efc7b5f5761a15fb6df23af0f1b0c94170bb106b3154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.432991 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.445281 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.455216 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.467550 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.475875 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.475913 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.475924 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.475938 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.475951 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:20Z","lastTransitionTime":"2025-10-01T16:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.479064 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:20Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.578671 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.578949 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.579079 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.579163 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.579222 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:20Z","lastTransitionTime":"2025-10-01T16:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.681957 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.682415 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.682606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.682742 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.682851 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:20Z","lastTransitionTime":"2025-10-01T16:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.785864 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.785895 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.785904 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.785917 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.785926 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:20Z","lastTransitionTime":"2025-10-01T16:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.807380 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:20 crc kubenswrapper[4726]: E1001 16:00:20.807566 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.807422 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:20 crc kubenswrapper[4726]: E1001 16:00:20.807797 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.807413 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:20 crc kubenswrapper[4726]: E1001 16:00:20.808038 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.888243 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.888298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.888314 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.888336 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.888353 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:20Z","lastTransitionTime":"2025-10-01T16:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.990798 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.990846 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.990858 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.990875 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:20 crc kubenswrapper[4726]: I1001 16:00:20.990888 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:20Z","lastTransitionTime":"2025-10-01T16:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.093178 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.093240 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.093266 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.093294 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.093318 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:21Z","lastTransitionTime":"2025-10-01T16:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.195265 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.195328 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.195350 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.195380 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.195404 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:21Z","lastTransitionTime":"2025-10-01T16:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.298142 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.298200 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.298218 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.298241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.298259 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:21Z","lastTransitionTime":"2025-10-01T16:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.400998 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.401092 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.401109 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.401132 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.401148 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:21Z","lastTransitionTime":"2025-10-01T16:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.503688 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.503962 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.504073 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.504217 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.504334 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:21Z","lastTransitionTime":"2025-10-01T16:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.606555 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.606830 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.606928 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.607019 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.607142 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:21Z","lastTransitionTime":"2025-10-01T16:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.710178 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.710464 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.710710 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.711182 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.711355 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:21Z","lastTransitionTime":"2025-10-01T16:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.807958 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:21 crc kubenswrapper[4726]: E1001 16:00:21.808451 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.814526 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.814584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.814605 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.814629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.814646 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:21Z","lastTransitionTime":"2025-10-01T16:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.916587 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.916664 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.916679 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.916701 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:21 crc kubenswrapper[4726]: I1001 16:00:21.916716 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:21Z","lastTransitionTime":"2025-10-01T16:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.019785 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.020030 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.020167 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.020262 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.020342 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:22Z","lastTransitionTime":"2025-10-01T16:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.123002 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.123111 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.123141 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.123168 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.123184 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:22Z","lastTransitionTime":"2025-10-01T16:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.226150 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.226217 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.226243 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.226271 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.226291 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:22Z","lastTransitionTime":"2025-10-01T16:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.321546 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.321624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.321646 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.321678 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.321701 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:22Z","lastTransitionTime":"2025-10-01T16:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:22 crc kubenswrapper[4726]: E1001 16:00:22.340158 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:22Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.345587 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.345645 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.345663 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.345686 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.345706 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:22Z","lastTransitionTime":"2025-10-01T16:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:22 crc kubenswrapper[4726]: E1001 16:00:22.361631 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:22Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.367648 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.367706 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.367725 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.367752 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.367773 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:22Z","lastTransitionTime":"2025-10-01T16:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:22 crc kubenswrapper[4726]: E1001 16:00:22.385874 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:22Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.390644 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.390754 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.390780 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.390813 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.390839 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:22Z","lastTransitionTime":"2025-10-01T16:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:22 crc kubenswrapper[4726]: E1001 16:00:22.406793 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:22Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.413005 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.413101 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.413119 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.413143 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.413160 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:22Z","lastTransitionTime":"2025-10-01T16:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:22 crc kubenswrapper[4726]: E1001 16:00:22.428144 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:22Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:22 crc kubenswrapper[4726]: E1001 16:00:22.428772 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.430774 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.430843 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.430869 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.430899 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.430923 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:22Z","lastTransitionTime":"2025-10-01T16:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.534515 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.534598 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.534648 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.534675 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.534693 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:22Z","lastTransitionTime":"2025-10-01T16:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.637579 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.637629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.637644 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.637662 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.637675 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:22Z","lastTransitionTime":"2025-10-01T16:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.740437 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.740477 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.740485 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.740500 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.740509 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:22Z","lastTransitionTime":"2025-10-01T16:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.807846 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.807916 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.807860 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:22 crc kubenswrapper[4726]: E1001 16:00:22.808091 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:22 crc kubenswrapper[4726]: E1001 16:00:22.808188 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:22 crc kubenswrapper[4726]: E1001 16:00:22.808288 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.843118 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.843174 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.843195 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.843219 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.843237 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:22Z","lastTransitionTime":"2025-10-01T16:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.945663 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.945732 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.945752 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.945778 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:22 crc kubenswrapper[4726]: I1001 16:00:22.945795 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:22Z","lastTransitionTime":"2025-10-01T16:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.048499 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.048544 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.048556 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.048574 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.048588 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:23Z","lastTransitionTime":"2025-10-01T16:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.152100 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.152145 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.152154 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.152173 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.152182 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:23Z","lastTransitionTime":"2025-10-01T16:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.254619 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.254709 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.254726 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.254745 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.254756 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:23Z","lastTransitionTime":"2025-10-01T16:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.359179 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.359246 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.359269 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.359299 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.359321 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:23Z","lastTransitionTime":"2025-10-01T16:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.462617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.462715 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.462737 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.462764 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.462837 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:23Z","lastTransitionTime":"2025-10-01T16:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.565453 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.565488 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.565500 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.565517 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.565533 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:23Z","lastTransitionTime":"2025-10-01T16:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.669164 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.669312 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.669338 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.669396 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.669425 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:23Z","lastTransitionTime":"2025-10-01T16:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.772086 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.772169 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.772206 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.772233 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.772251 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:23Z","lastTransitionTime":"2025-10-01T16:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.808027 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:23 crc kubenswrapper[4726]: E1001 16:00:23.808466 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.875327 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.875652 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.875796 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.875980 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.876187 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:23Z","lastTransitionTime":"2025-10-01T16:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.979847 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.979920 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.979932 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.979948 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:23 crc kubenswrapper[4726]: I1001 16:00:23.979959 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:23Z","lastTransitionTime":"2025-10-01T16:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.083170 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.083229 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.083245 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.083265 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.083277 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:24Z","lastTransitionTime":"2025-10-01T16:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.185139 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.185221 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.185259 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.185295 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.185317 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:24Z","lastTransitionTime":"2025-10-01T16:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.287953 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.287996 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.288004 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.288020 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.288030 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:24Z","lastTransitionTime":"2025-10-01T16:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.390795 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.390832 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.390844 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.390859 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.390870 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:24Z","lastTransitionTime":"2025-10-01T16:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.493124 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.493396 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.493536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.493650 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.493829 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:24Z","lastTransitionTime":"2025-10-01T16:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.596610 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.596691 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.596711 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.596754 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.596784 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:24Z","lastTransitionTime":"2025-10-01T16:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.700136 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.700193 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.700208 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.700231 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.700246 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:24Z","lastTransitionTime":"2025-10-01T16:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.802847 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.802887 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.802901 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.802917 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.802929 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:24Z","lastTransitionTime":"2025-10-01T16:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.807453 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.807472 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:24 crc kubenswrapper[4726]: E1001 16:00:24.807535 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:24 crc kubenswrapper[4726]: E1001 16:00:24.807660 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.807468 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:24 crc kubenswrapper[4726]: E1001 16:00:24.808207 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.906514 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.906567 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.906584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.906608 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:24 crc kubenswrapper[4726]: I1001 16:00:24.906625 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:24Z","lastTransitionTime":"2025-10-01T16:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.009153 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.010008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.010250 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.010472 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.010648 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:25Z","lastTransitionTime":"2025-10-01T16:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.114077 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.114116 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.114130 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.114147 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.114159 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:25Z","lastTransitionTime":"2025-10-01T16:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.217074 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.217144 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.217167 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.217196 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.217253 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:25Z","lastTransitionTime":"2025-10-01T16:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.320208 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.320296 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.320342 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.320364 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.320381 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:25Z","lastTransitionTime":"2025-10-01T16:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.423313 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.423389 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.423403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.423420 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.423431 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:25Z","lastTransitionTime":"2025-10-01T16:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.526540 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.526605 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.526621 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.526641 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.526661 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:25Z","lastTransitionTime":"2025-10-01T16:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.629504 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.629537 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.629545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.629558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.629568 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:25Z","lastTransitionTime":"2025-10-01T16:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.732473 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.732536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.732552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.732575 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.732594 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:25Z","lastTransitionTime":"2025-10-01T16:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.807039 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:25 crc kubenswrapper[4726]: E1001 16:00:25.807266 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.835605 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.835685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.835710 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.835742 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.835765 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:25Z","lastTransitionTime":"2025-10-01T16:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.938531 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.938568 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.938577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.938591 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:25 crc kubenswrapper[4726]: I1001 16:00:25.938601 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:25Z","lastTransitionTime":"2025-10-01T16:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.041528 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.041590 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.041606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.041636 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.041653 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:26Z","lastTransitionTime":"2025-10-01T16:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.144298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.144354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.144367 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.144382 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.144393 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:26Z","lastTransitionTime":"2025-10-01T16:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.246590 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.246651 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.246666 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.246692 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.246705 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:26Z","lastTransitionTime":"2025-10-01T16:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.349552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.349679 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.349696 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.349720 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.349737 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:26Z","lastTransitionTime":"2025-10-01T16:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.453044 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.453139 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.453156 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.453182 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.453200 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:26Z","lastTransitionTime":"2025-10-01T16:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.556042 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.556136 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.556156 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.556180 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.556198 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:26Z","lastTransitionTime":"2025-10-01T16:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.659457 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.659773 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.659881 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.659972 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.660152 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:26Z","lastTransitionTime":"2025-10-01T16:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.763354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.763419 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.763442 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.763472 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.763495 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:26Z","lastTransitionTime":"2025-10-01T16:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.807356 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.807370 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.807492 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:26 crc kubenswrapper[4726]: E1001 16:00:26.808018 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:26 crc kubenswrapper[4726]: E1001 16:00:26.808111 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:26 crc kubenswrapper[4726]: E1001 16:00:26.807825 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.866581 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.866634 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.866646 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.866661 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.866671 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:26Z","lastTransitionTime":"2025-10-01T16:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.969390 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.969436 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.969448 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.969466 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:26 crc kubenswrapper[4726]: I1001 16:00:26.969477 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:26Z","lastTransitionTime":"2025-10-01T16:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.072272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.072304 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.072314 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.072326 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.072335 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:27Z","lastTransitionTime":"2025-10-01T16:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.174842 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.174919 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.174946 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.174978 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.174997 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:27Z","lastTransitionTime":"2025-10-01T16:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.278018 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.278101 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.278115 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.278137 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.278151 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:27Z","lastTransitionTime":"2025-10-01T16:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.381265 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.381305 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.381315 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.381330 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.381342 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:27Z","lastTransitionTime":"2025-10-01T16:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.483777 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.483819 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.483828 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.483842 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.483851 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:27Z","lastTransitionTime":"2025-10-01T16:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.586677 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.587009 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.587027 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.587089 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.587115 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:27Z","lastTransitionTime":"2025-10-01T16:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.689839 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.689907 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.689930 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.689958 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.689977 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:27Z","lastTransitionTime":"2025-10-01T16:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.793158 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.793211 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.793231 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.793254 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.793271 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:27Z","lastTransitionTime":"2025-10-01T16:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.807047 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:27 crc kubenswrapper[4726]: E1001 16:00:27.807225 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.824630 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:27Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.842718 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:27Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.855968 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:27Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.867148 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:27Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.879997 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:27Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.896950 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:27Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.897211 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.897236 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.897247 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.897265 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.897276 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:27Z","lastTransitionTime":"2025-10-01T16:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.910552 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3064c56-dd95-414e-83a1-8876f73866fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4179687cd0b6fa600a3efc7b5f5761a15fb6df23af0f1b0c94170bb106b3154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:27Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.928874 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:27Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.943777 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12eb575b-72e7-4bb2-b9de-ce924d47db1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9fe86b3d6a1acd9512663da493ea130d4112b79d2ea24e5d94eeb76e81d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f94a8081041e5968c831dce9988a771386310ae721226c4f58baa6a2c1af47b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565eeb4bdc62a6f23f5a056700c8677e7de8bac37bd7cc766d4b8637c7fe3f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:27Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.958169 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:27Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.972625 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:27Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.985936 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e127584cf73c7f9ea20da97dce5f50fb86b8f9fde6ae383172834d425bb1e033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:18Z\\\",\\\"message\\\":\\\"2025-10-01T15:59:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14\\\\n2025-10-01T15:59:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14 to /host/opt/cni/bin/\\\\n2025-10-01T15:59:33Z [verbose] multus-daemon started\\\\n2025-10-01T15:59:33Z [verbose] Readiness Indicator file check\\\\n2025-10-01T16:00:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T16:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:27Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:27 crc kubenswrapper[4726]: I1001 16:00:27.999190 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:27Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.000735 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.000763 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.000774 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.000789 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.000799 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:28Z","lastTransitionTime":"2025-10-01T16:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.023842 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:02Z\\\",\\\"message\\\":\\\"oller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z]\\\\nI1001 16:00:02.627548 6407 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager-operator/metrics for network=default are: map[]\\\\nI1001 16:00:02.627560 6407 services_controller.go:443] Built service openshift-kube-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.219\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1001 16:00:02.627572 6407 services_controller.go:444] Built service opens\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T16:00:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:28Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.036475 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:28Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.050297 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:28Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.068289 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:28Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.082271 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:28Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.103625 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.103671 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.103682 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.103700 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.103712 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:28Z","lastTransitionTime":"2025-10-01T16:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.206122 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.206162 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.206172 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.206185 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.206193 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:28Z","lastTransitionTime":"2025-10-01T16:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.308469 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.308565 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.308585 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.308648 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.308668 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:28Z","lastTransitionTime":"2025-10-01T16:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.411542 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.411668 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.411692 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.411717 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.411734 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:28Z","lastTransitionTime":"2025-10-01T16:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.514656 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.514725 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.514745 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.514776 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.514795 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:28Z","lastTransitionTime":"2025-10-01T16:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.617489 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.617781 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.617868 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.617975 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.618101 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:28Z","lastTransitionTime":"2025-10-01T16:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.720293 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.720346 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.720361 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.720380 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.720399 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:28Z","lastTransitionTime":"2025-10-01T16:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.807618 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.807698 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.807620 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:28 crc kubenswrapper[4726]: E1001 16:00:28.807829 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:28 crc kubenswrapper[4726]: E1001 16:00:28.808024 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:28 crc kubenswrapper[4726]: E1001 16:00:28.808225 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.822818 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.822868 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.822916 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.822940 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.822957 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:28Z","lastTransitionTime":"2025-10-01T16:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.926081 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.926180 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.926200 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.926253 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:28 crc kubenswrapper[4726]: I1001 16:00:28.926280 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:28Z","lastTransitionTime":"2025-10-01T16:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.028988 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.029078 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.029089 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.029104 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.029113 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:29Z","lastTransitionTime":"2025-10-01T16:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.133044 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.133750 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.133778 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.133987 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.134019 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:29Z","lastTransitionTime":"2025-10-01T16:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.238175 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.238232 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.238249 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.238273 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.238293 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:29Z","lastTransitionTime":"2025-10-01T16:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.341260 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.341310 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.341350 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.341370 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.341383 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:29Z","lastTransitionTime":"2025-10-01T16:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.444565 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.444634 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.444647 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.444664 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.444673 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:29Z","lastTransitionTime":"2025-10-01T16:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.548440 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.548520 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.548547 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.548577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.548601 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:29Z","lastTransitionTime":"2025-10-01T16:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.652163 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.652225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.652241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.652264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.652285 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:29Z","lastTransitionTime":"2025-10-01T16:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.756140 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.756201 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.756227 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.756273 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.756296 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:29Z","lastTransitionTime":"2025-10-01T16:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.808100 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:29 crc kubenswrapper[4726]: E1001 16:00:29.808293 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.859102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.859142 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.859155 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.859172 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.859187 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:29Z","lastTransitionTime":"2025-10-01T16:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.961761 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.961808 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.961825 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.961849 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:29 crc kubenswrapper[4726]: I1001 16:00:29.961967 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:29Z","lastTransitionTime":"2025-10-01T16:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.064768 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.064822 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.064838 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.064861 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.064877 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:30Z","lastTransitionTime":"2025-10-01T16:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.167969 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.168006 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.168017 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.168033 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.168067 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:30Z","lastTransitionTime":"2025-10-01T16:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.271273 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.271334 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.271353 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.271379 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.271412 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:30Z","lastTransitionTime":"2025-10-01T16:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.374598 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.374659 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.374676 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.374701 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.374719 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:30Z","lastTransitionTime":"2025-10-01T16:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.477769 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.477862 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.477925 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.477951 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.477969 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:30Z","lastTransitionTime":"2025-10-01T16:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.582199 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.582267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.582287 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.582317 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.582337 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:30Z","lastTransitionTime":"2025-10-01T16:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.685485 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.685551 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.685562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.685578 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.685587 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:30Z","lastTransitionTime":"2025-10-01T16:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.788802 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.789345 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.789524 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.789707 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.789916 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:30Z","lastTransitionTime":"2025-10-01T16:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.807319 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.807330 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:30 crc kubenswrapper[4726]: E1001 16:00:30.807500 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:30 crc kubenswrapper[4726]: E1001 16:00:30.807602 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.807746 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:30 crc kubenswrapper[4726]: E1001 16:00:30.807995 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.892170 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.892200 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.892211 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.892226 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.892236 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:30Z","lastTransitionTime":"2025-10-01T16:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.995140 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.995184 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.995194 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.995210 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:30 crc kubenswrapper[4726]: I1001 16:00:30.995219 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:30Z","lastTransitionTime":"2025-10-01T16:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.098028 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.098108 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.098122 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.098151 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.098173 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:31Z","lastTransitionTime":"2025-10-01T16:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.200945 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.201020 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.201044 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.201100 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.201124 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:31Z","lastTransitionTime":"2025-10-01T16:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.304199 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.304269 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.304292 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.304321 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.304342 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:31Z","lastTransitionTime":"2025-10-01T16:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.406984 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.407037 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.407083 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.407107 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.407125 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:31Z","lastTransitionTime":"2025-10-01T16:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.510340 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.510397 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.510415 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.510444 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.510467 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:31Z","lastTransitionTime":"2025-10-01T16:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.613348 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.613439 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.613460 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.613485 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.613502 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:31Z","lastTransitionTime":"2025-10-01T16:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.717238 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.717362 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.717401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.717434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.717460 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:31Z","lastTransitionTime":"2025-10-01T16:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.808044 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:31 crc kubenswrapper[4726]: E1001 16:00:31.808290 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.821034 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.821236 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.821273 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.821304 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.821326 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:31Z","lastTransitionTime":"2025-10-01T16:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.924138 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.924194 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.924212 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.924235 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:31 crc kubenswrapper[4726]: I1001 16:00:31.924253 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:31Z","lastTransitionTime":"2025-10-01T16:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.027281 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.027349 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.027367 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.027394 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.027412 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:32Z","lastTransitionTime":"2025-10-01T16:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.130719 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.131160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.131401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.131637 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.131851 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:32Z","lastTransitionTime":"2025-10-01T16:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.235044 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.235107 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.235117 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.235133 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.235146 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:32Z","lastTransitionTime":"2025-10-01T16:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.339632 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.339684 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.339701 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.339723 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.339740 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:32Z","lastTransitionTime":"2025-10-01T16:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.442018 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.442093 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.442105 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.442123 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.442138 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:32Z","lastTransitionTime":"2025-10-01T16:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.544957 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.545046 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.545106 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.545136 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.545157 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:32Z","lastTransitionTime":"2025-10-01T16:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.607502 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.607590 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.607612 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.607644 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.607664 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:32Z","lastTransitionTime":"2025-10-01T16:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:32 crc kubenswrapper[4726]: E1001 16:00:32.626172 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:32Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.631206 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.631264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.631283 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.631308 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.631326 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:32Z","lastTransitionTime":"2025-10-01T16:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:32 crc kubenswrapper[4726]: E1001 16:00:32.647662 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:32Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.653008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.653089 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.653108 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.653134 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.653151 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:32Z","lastTransitionTime":"2025-10-01T16:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:32 crc kubenswrapper[4726]: E1001 16:00:32.668886 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:32Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.674432 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.674613 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.674773 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.674898 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.675037 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:32Z","lastTransitionTime":"2025-10-01T16:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:32 crc kubenswrapper[4726]: E1001 16:00:32.690514 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:32Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.694371 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.694437 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.694450 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.694464 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.694475 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:32Z","lastTransitionTime":"2025-10-01T16:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:32 crc kubenswrapper[4726]: E1001 16:00:32.707988 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:32Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:32 crc kubenswrapper[4726]: E1001 16:00:32.708177 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.710128 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.710175 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.710191 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.710213 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.710237 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:32Z","lastTransitionTime":"2025-10-01T16:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.807366 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.807416 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.807366 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:32 crc kubenswrapper[4726]: E1001 16:00:32.807587 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:32 crc kubenswrapper[4726]: E1001 16:00:32.807643 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:32 crc kubenswrapper[4726]: E1001 16:00:32.807702 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.813102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.813162 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.813185 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.813214 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.813281 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:32Z","lastTransitionTime":"2025-10-01T16:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.916008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.916122 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.916148 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.916617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:32 crc kubenswrapper[4726]: I1001 16:00:32.916892 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:32Z","lastTransitionTime":"2025-10-01T16:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.020654 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.020721 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.020744 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.020773 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.020796 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:33Z","lastTransitionTime":"2025-10-01T16:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.124882 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.124967 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.124998 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.125029 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.125085 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:33Z","lastTransitionTime":"2025-10-01T16:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.228018 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.228119 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.228144 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.228173 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.228195 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:33Z","lastTransitionTime":"2025-10-01T16:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.331764 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.331829 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.331848 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.331872 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.331889 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:33Z","lastTransitionTime":"2025-10-01T16:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.434346 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.434376 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.434384 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.434398 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.434407 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:33Z","lastTransitionTime":"2025-10-01T16:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.537014 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.537043 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.537083 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.537098 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.537107 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:33Z","lastTransitionTime":"2025-10-01T16:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.640471 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.640538 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.640555 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.640581 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.640601 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:33Z","lastTransitionTime":"2025-10-01T16:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.672356 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.672567 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:33 crc kubenswrapper[4726]: E1001 16:00:33.672608 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.672571945 +0000 UTC m=+150.574124532 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:00:33 crc kubenswrapper[4726]: E1001 16:00:33.672739 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 16:00:33 crc kubenswrapper[4726]: E1001 16:00:33.672765 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 16:00:33 crc kubenswrapper[4726]: E1001 16:00:33.672779 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 16:00:33 crc kubenswrapper[4726]: E1001 16:00:33.672835 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.672819722 +0000 UTC m=+150.574372309 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.743270 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.743358 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.743382 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.743411 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.743433 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:33Z","lastTransitionTime":"2025-10-01T16:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.774099 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.774170 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.774198 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:33 crc kubenswrapper[4726]: E1001 16:00:33.774272 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 16:00:33 crc kubenswrapper[4726]: E1001 16:00:33.774325 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.774309622 +0000 UTC m=+150.675862209 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 16:00:33 crc kubenswrapper[4726]: E1001 16:00:33.774326 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 16:00:33 crc kubenswrapper[4726]: E1001 16:00:33.774368 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 16:00:33 crc kubenswrapper[4726]: E1001 16:00:33.774409 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 16:00:33 crc kubenswrapper[4726]: E1001 16:00:33.774435 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 16:00:33 crc kubenswrapper[4726]: E1001 16:00:33.774379 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.774367534 +0000 UTC m=+150.675920111 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 16:00:33 crc kubenswrapper[4726]: E1001 16:00:33.774534 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.774509067 +0000 UTC m=+150.676061684 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.807513 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:33 crc kubenswrapper[4726]: E1001 16:00:33.807659 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.808766 4726 scope.go:117] "RemoveContainer" containerID="63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.845940 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.845974 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.845983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.845998 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.846010 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:33Z","lastTransitionTime":"2025-10-01T16:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.949290 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.949354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.949371 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.949395 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:33 crc kubenswrapper[4726]: I1001 16:00:33.949417 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:33Z","lastTransitionTime":"2025-10-01T16:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.051894 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.051932 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.051942 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.051956 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.051965 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:34Z","lastTransitionTime":"2025-10-01T16:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.154081 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.154131 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.154143 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.154164 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.154176 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:34Z","lastTransitionTime":"2025-10-01T16:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.256657 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.256694 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.256704 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.256718 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.256729 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:34Z","lastTransitionTime":"2025-10-01T16:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.308557 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/2.log" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.311250 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerStarted","Data":"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be"} Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.312089 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.335147 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.347040 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.357991 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.358976 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.359002 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.359010 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.359024 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.359033 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:34Z","lastTransitionTime":"2025-10-01T16:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.374413 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:02Z\\\",\\\"message\\\":\\\"oller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z]\\\\nI1001 16:00:02.627548 6407 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager-operator/metrics for network=default are: map[]\\\\nI1001 16:00:02.627560 6407 services_controller.go:443] Built service openshift-kube-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.219\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1001 16:00:02.627572 6407 services_controller.go:444] Built service opens\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T16:00:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T16:00:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.386322 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3064c56-dd95-414e-83a1-8876f73866fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4179687cd0b6fa600a3efc7b5f5761a15fb6df23af0f1b0c94170bb106b3154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.406419 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.419144 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12eb575b-72e7-4bb2-b9de-ce924d47db1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9fe86b3d6a1acd9512663da493ea130d4112b79d2ea24e5d94eeb76e81d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f94a8081041e5968c831dce9988a771386310ae721226c4f58baa6a2c1af47b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565eeb4bdc62a6f23f5a056700c8677e7de8bac37bd7cc766d4b8637c7fe3f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.432307 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.452020 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.461406 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.461440 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.461449 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.461461 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.461470 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:34Z","lastTransitionTime":"2025-10-01T16:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.473204 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e127584cf73c7f9ea20da97dce5f50fb86b8f9fde6ae383172834d425bb1e033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:18Z\\\",\\\"message\\\":\\\"2025-10-01T15:59:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14\\\\n2025-10-01T15:59:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14 to /host/opt/cni/bin/\\\\n2025-10-01T15:59:33Z [verbose] multus-daemon started\\\\n2025-10-01T15:59:33Z [verbose] Readiness Indicator file check\\\\n2025-10-01T16:00:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T16:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.484672 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.495862 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.509069 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.517862 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.528978 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.542341 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.555482 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.563572 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.563612 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.563628 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.563645 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.563656 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:34Z","lastTransitionTime":"2025-10-01T16:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.567280 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:34Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.666486 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.666714 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.666795 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.666882 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.666955 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:34Z","lastTransitionTime":"2025-10-01T16:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.769076 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.769121 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.769136 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.769160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.769177 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:34Z","lastTransitionTime":"2025-10-01T16:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.806986 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.807120 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:34 crc kubenswrapper[4726]: E1001 16:00:34.807184 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:34 crc kubenswrapper[4726]: E1001 16:00:34.807364 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.807572 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:34 crc kubenswrapper[4726]: E1001 16:00:34.807891 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.871347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.871401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.871417 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.871439 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.871458 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:34Z","lastTransitionTime":"2025-10-01T16:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.975619 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.975714 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.975740 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.975777 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:34 crc kubenswrapper[4726]: I1001 16:00:34.975797 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:34Z","lastTransitionTime":"2025-10-01T16:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.080490 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.080858 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.080874 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.080898 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.080915 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:35Z","lastTransitionTime":"2025-10-01T16:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.183745 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.183821 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.183845 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.183874 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.183894 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:35Z","lastTransitionTime":"2025-10-01T16:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.287363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.287428 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.287445 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.287468 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.287485 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:35Z","lastTransitionTime":"2025-10-01T16:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.318092 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/3.log" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.319392 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/2.log" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.322976 4726 generic.go:334] "Generic (PLEG): container finished" podID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerID="04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be" exitCode=1 Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.323045 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerDied","Data":"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be"} Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.323100 4726 scope.go:117] "RemoveContainer" containerID="63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.324390 4726 scope.go:117] "RemoveContainer" containerID="04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be" Oct 01 16:00:35 crc kubenswrapper[4726]: E1001 16:00:35.324668 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.341140 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.363523 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.383145 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.390010 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.390153 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.390185 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.390217 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.390249 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:35Z","lastTransitionTime":"2025-10-01T16:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.419785 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63eaa52aa3aafd88e1f167442d3a1b9dc3c6aa1d33079dbd93c4fcfa61f86be9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:02Z\\\",\\\"message\\\":\\\"oller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:02Z is after 2025-08-24T17:21:41Z]\\\\nI1001 16:00:02.627548 6407 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager-operator/metrics for network=default are: map[]\\\\nI1001 16:00:02.627560 6407 services_controller.go:443] Built service openshift-kube-controller-manager-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.219\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1001 16:00:02.627572 6407 services_controller.go:444] Built service opens\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T16:00:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:34Z\\\",\\\"message\\\":\\\"ssqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 16:00:34.715200 6847 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 16:00:34.715257 6847 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 16:00:34.715288 6847 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 16:00:34.715698 6847 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 16:00:34.715725 6847 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 16:00:34.715741 6847 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 16:00:34.715746 6847 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 16:00:34.715818 6847 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 16:00:34.715835 6847 factory.go:656] Stopping watch factory\\\\nI1001 16:00:34.715835 6847 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 16:00:34.715846 6847 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 16:00:34.715859 6847 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T16:00:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.437327 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3064c56-dd95-414e-83a1-8876f73866fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4179687cd0b6fa600a3efc7b5f5761a15fb6df23af0f1b0c94170bb106b3154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.455245 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.467937 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12eb575b-72e7-4bb2-b9de-ce924d47db1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9fe86b3d6a1acd9512663da493ea130d4112b79d2ea24e5d94eeb76e81d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f94a8081041e5968c831dce9988a771386310ae721226c4f58baa6a2c1af47b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565eeb4bdc62a6f23f5a056700c8677e7de8bac37bd7cc766d4b8637c7fe3f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.481626 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.492718 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.492749 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.492760 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.492776 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.492788 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:35Z","lastTransitionTime":"2025-10-01T16:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.501578 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.518403 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e127584cf73c7f9ea20da97dce5f50fb86b8f9fde6ae383172834d425bb1e033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:18Z\\\",\\\"message\\\":\\\"2025-10-01T15:59:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14\\\\n2025-10-01T15:59:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14 to /host/opt/cni/bin/\\\\n2025-10-01T15:59:33Z [verbose] multus-daemon started\\\\n2025-10-01T15:59:33Z [verbose] Readiness Indicator file check\\\\n2025-10-01T16:00:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T16:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.535450 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.545828 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.559582 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.573907 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.589345 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.594909 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.594954 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.594965 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.594981 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.594994 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:35Z","lastTransitionTime":"2025-10-01T16:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.608720 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.625838 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.637967 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:35Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.698282 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.698354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.698372 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.698397 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.698414 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:35Z","lastTransitionTime":"2025-10-01T16:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.801289 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.801367 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.801387 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.801412 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.801429 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:35Z","lastTransitionTime":"2025-10-01T16:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.807968 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:35 crc kubenswrapper[4726]: E1001 16:00:35.808209 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.903972 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.904022 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.904033 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.904079 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:35 crc kubenswrapper[4726]: I1001 16:00:35.904095 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:35Z","lastTransitionTime":"2025-10-01T16:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.006541 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.006581 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.006596 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.006611 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.006648 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:36Z","lastTransitionTime":"2025-10-01T16:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.109753 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.109824 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.109846 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.109875 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.109900 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:36Z","lastTransitionTime":"2025-10-01T16:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.213026 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.213137 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.213163 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.213193 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.213217 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:36Z","lastTransitionTime":"2025-10-01T16:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.316009 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.316127 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.316146 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.316169 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.316187 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:36Z","lastTransitionTime":"2025-10-01T16:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.328442 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/3.log" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.333725 4726 scope.go:117] "RemoveContainer" containerID="04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be" Oct 01 16:00:36 crc kubenswrapper[4726]: E1001 16:00:36.334013 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.354093 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e127584cf73c7f9ea20da97dce5f50fb86b8f9fde6ae383172834d425bb1e033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:18Z\\\",\\\"message\\\":\\\"2025-10-01T15:59:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14\\\\n2025-10-01T15:59:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14 to /host/opt/cni/bin/\\\\n2025-10-01T15:59:33Z [verbose] multus-daemon started\\\\n2025-10-01T15:59:33Z [verbose] Readiness Indicator file check\\\\n2025-10-01T16:00:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T16:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.373926 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.407768 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:34Z\\\",\\\"message\\\":\\\"ssqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 16:00:34.715200 6847 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 16:00:34.715257 6847 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 16:00:34.715288 6847 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 16:00:34.715698 6847 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 16:00:34.715725 6847 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 16:00:34.715741 6847 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 16:00:34.715746 6847 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 16:00:34.715818 6847 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 16:00:34.715835 6847 factory.go:656] Stopping watch factory\\\\nI1001 16:00:34.715835 6847 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 16:00:34.715846 6847 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 16:00:34.715859 6847 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T16:00:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.419612 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.419679 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.419697 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.419721 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.419742 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:36Z","lastTransitionTime":"2025-10-01T16:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.426452 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3064c56-dd95-414e-83a1-8876f73866fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4179687cd0b6fa600a3efc7b5f5761a15fb6df23af0f1b0c94170bb106b3154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.449447 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.466393 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12eb575b-72e7-4bb2-b9de-ce924d47db1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9fe86b3d6a1acd9512663da493ea130d4112b79d2ea24e5d94eeb76e81d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f94a8081041e5968c831dce9988a771386310ae721226c4f58baa6a2c1af47b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565eeb4bdc62a6f23f5a056700c8677e7de8bac37bd7cc766d4b8637c7fe3f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.486772 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.502459 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.518258 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.522222 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.522257 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.522268 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.522286 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.522299 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:36Z","lastTransitionTime":"2025-10-01T16:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.534910 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.551120 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.562525 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.574472 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.592376 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.610525 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.622596 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.624371 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.624422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.624438 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.624460 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.624475 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:36Z","lastTransitionTime":"2025-10-01T16:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.635167 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.647109 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:36Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.727193 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.727264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.727294 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.727322 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.727343 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:36Z","lastTransitionTime":"2025-10-01T16:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.807192 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.807197 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:36 crc kubenswrapper[4726]: E1001 16:00:36.807362 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.807514 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:36 crc kubenswrapper[4726]: E1001 16:00:36.807582 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:36 crc kubenswrapper[4726]: E1001 16:00:36.807709 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.830134 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.830313 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.830336 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.830359 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.830376 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:36Z","lastTransitionTime":"2025-10-01T16:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.934580 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.934653 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.934695 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.934728 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:36 crc kubenswrapper[4726]: I1001 16:00:36.934750 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:36Z","lastTransitionTime":"2025-10-01T16:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.038333 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.038402 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.038421 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.038446 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.038465 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:37Z","lastTransitionTime":"2025-10-01T16:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.140712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.140776 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.140794 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.140817 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.140836 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:37Z","lastTransitionTime":"2025-10-01T16:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.244146 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.244205 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.244222 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.244244 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.244264 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:37Z","lastTransitionTime":"2025-10-01T16:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.346691 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.346766 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.346786 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.346813 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.346833 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:37Z","lastTransitionTime":"2025-10-01T16:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.449238 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.449295 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.449312 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.449333 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.449349 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:37Z","lastTransitionTime":"2025-10-01T16:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.551960 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.552225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.552250 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.552275 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.552292 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:37Z","lastTransitionTime":"2025-10-01T16:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.656231 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.656301 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.656338 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.656369 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.656390 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:37Z","lastTransitionTime":"2025-10-01T16:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.759326 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.759398 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.759417 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.759445 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.759465 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:37Z","lastTransitionTime":"2025-10-01T16:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.807356 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:37 crc kubenswrapper[4726]: E1001 16:00:37.807526 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.822043 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:37Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.839964 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:37Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.856999 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12eb575b-72e7-4bb2-b9de-ce924d47db1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9fe86b3d6a1acd9512663da493ea130d4112b79d2ea24e5d94eeb76e81d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f94a8081041e5968c831dce9988a771386310ae721226c4f58baa6a2c1af47b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565eeb4bdc62a6f23f5a056700c8677e7de8bac37bd7cc766d4b8637c7fe3f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:37Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.862291 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.862338 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.862355 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.862378 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.862396 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:37Z","lastTransitionTime":"2025-10-01T16:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.878013 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:37Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.892202 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:37Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.914568 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e127584cf73c7f9ea20da97dce5f50fb86b8f9fde6ae383172834d425bb1e033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:18Z\\\",\\\"message\\\":\\\"2025-10-01T15:59:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14\\\\n2025-10-01T15:59:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14 to /host/opt/cni/bin/\\\\n2025-10-01T15:59:33Z [verbose] multus-daemon started\\\\n2025-10-01T15:59:33Z [verbose] Readiness Indicator file check\\\\n2025-10-01T16:00:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T16:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:37Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.934321 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:37Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.965342 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.965411 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.965437 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.965464 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.965482 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:37Z","lastTransitionTime":"2025-10-01T16:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.967935 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:34Z\\\",\\\"message\\\":\\\"ssqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 16:00:34.715200 6847 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 16:00:34.715257 6847 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 16:00:34.715288 6847 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 16:00:34.715698 6847 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 16:00:34.715725 6847 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 16:00:34.715741 6847 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 16:00:34.715746 6847 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 16:00:34.715818 6847 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 16:00:34.715835 6847 factory.go:656] Stopping watch factory\\\\nI1001 16:00:34.715835 6847 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 16:00:34.715846 6847 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 16:00:34.715859 6847 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T16:00:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:37Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:37 crc kubenswrapper[4726]: I1001 16:00:37.985363 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3064c56-dd95-414e-83a1-8876f73866fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4179687cd0b6fa600a3efc7b5f5761a15fb6df23af0f1b0c94170bb106b3154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:37Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.009597 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:38Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.032305 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:38Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.049201 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:38Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.069036 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.069123 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.069144 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.069167 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.069185 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:38Z","lastTransitionTime":"2025-10-01T16:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.069608 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:38Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.087792 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:38Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.108895 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:38Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.126612 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:38Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.143404 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:38Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.164897 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:38Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.171565 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.171619 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.171636 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.171661 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.171677 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:38Z","lastTransitionTime":"2025-10-01T16:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.274516 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.274566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.274667 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.274736 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.274783 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:38Z","lastTransitionTime":"2025-10-01T16:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.377999 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.378100 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.378121 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.378143 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.378161 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:38Z","lastTransitionTime":"2025-10-01T16:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.481354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.481421 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.481462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.481503 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.481525 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:38Z","lastTransitionTime":"2025-10-01T16:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.584515 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.584587 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.584611 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.584640 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.584662 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:38Z","lastTransitionTime":"2025-10-01T16:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.688193 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.688255 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.688274 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.688297 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.688314 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:38Z","lastTransitionTime":"2025-10-01T16:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.791449 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.791487 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.791500 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.791524 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.791538 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:38Z","lastTransitionTime":"2025-10-01T16:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.807397 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:38 crc kubenswrapper[4726]: E1001 16:00:38.807533 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.807398 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.807398 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:38 crc kubenswrapper[4726]: E1001 16:00:38.807606 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:38 crc kubenswrapper[4726]: E1001 16:00:38.807765 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.894638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.894710 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.894750 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.894782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.894806 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:38Z","lastTransitionTime":"2025-10-01T16:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.997729 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.997803 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.997817 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.997843 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:38 crc kubenswrapper[4726]: I1001 16:00:38.997861 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:38Z","lastTransitionTime":"2025-10-01T16:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.100974 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.101039 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.101073 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.101095 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.101108 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:39Z","lastTransitionTime":"2025-10-01T16:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.204790 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.204902 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.204921 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.204948 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.204970 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:39Z","lastTransitionTime":"2025-10-01T16:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.308720 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.308794 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.308811 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.308830 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.308843 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:39Z","lastTransitionTime":"2025-10-01T16:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.412008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.412061 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.412071 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.412085 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.412095 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:39Z","lastTransitionTime":"2025-10-01T16:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.515300 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.515359 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.515377 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.515402 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.515421 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:39Z","lastTransitionTime":"2025-10-01T16:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.617934 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.617997 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.618018 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.618043 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.618095 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:39Z","lastTransitionTime":"2025-10-01T16:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.720357 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.720398 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.720411 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.720430 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.720445 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:39Z","lastTransitionTime":"2025-10-01T16:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.807104 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:39 crc kubenswrapper[4726]: E1001 16:00:39.807255 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.823723 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.823806 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.823830 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.823859 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.823886 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:39Z","lastTransitionTime":"2025-10-01T16:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.926402 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.926480 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.926499 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.926527 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:39 crc kubenswrapper[4726]: I1001 16:00:39.926547 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:39Z","lastTransitionTime":"2025-10-01T16:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.030448 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.030491 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.030500 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.030531 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.030543 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:40Z","lastTransitionTime":"2025-10-01T16:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.134251 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.134296 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.134309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.134324 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.134335 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:40Z","lastTransitionTime":"2025-10-01T16:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.237156 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.237208 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.237221 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.237241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.237259 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:40Z","lastTransitionTime":"2025-10-01T16:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.340927 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.340985 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.341006 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.341029 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.341076 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:40Z","lastTransitionTime":"2025-10-01T16:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.444274 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.444334 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.444356 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.444382 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.444400 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:40Z","lastTransitionTime":"2025-10-01T16:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.547609 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.547689 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.547714 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.547742 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.547764 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:40Z","lastTransitionTime":"2025-10-01T16:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.651482 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.651536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.651554 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.651580 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.651598 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:40Z","lastTransitionTime":"2025-10-01T16:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.757996 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.758096 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.758124 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.758143 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.758156 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:40Z","lastTransitionTime":"2025-10-01T16:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.807732 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.807806 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.807880 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:40 crc kubenswrapper[4726]: E1001 16:00:40.807992 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:40 crc kubenswrapper[4726]: E1001 16:00:40.808382 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:40 crc kubenswrapper[4726]: E1001 16:00:40.808472 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.861549 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.861630 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.861656 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.861686 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.861707 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:40Z","lastTransitionTime":"2025-10-01T16:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.964771 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.964818 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.964830 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.964847 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:40 crc kubenswrapper[4726]: I1001 16:00:40.964875 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:40Z","lastTransitionTime":"2025-10-01T16:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.067518 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.067609 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.067638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.067672 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.067695 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:41Z","lastTransitionTime":"2025-10-01T16:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.170734 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.170787 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.170796 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.170811 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.170823 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:41Z","lastTransitionTime":"2025-10-01T16:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.274257 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.274351 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.274383 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.274413 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.274431 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:41Z","lastTransitionTime":"2025-10-01T16:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.377709 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.377762 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.377775 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.377792 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.377807 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:41Z","lastTransitionTime":"2025-10-01T16:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.481129 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.481201 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.481226 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.481256 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.481278 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:41Z","lastTransitionTime":"2025-10-01T16:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.584250 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.584373 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.584395 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.584419 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.584436 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:41Z","lastTransitionTime":"2025-10-01T16:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.687183 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.687266 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.687287 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.687315 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.687337 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:41Z","lastTransitionTime":"2025-10-01T16:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.790322 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.790395 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.790412 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.790436 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.790453 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:41Z","lastTransitionTime":"2025-10-01T16:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.807577 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:41 crc kubenswrapper[4726]: E1001 16:00:41.807780 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.827248 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.894514 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.894568 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.894585 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.894608 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.894626 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:41Z","lastTransitionTime":"2025-10-01T16:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.997187 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.997247 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.997263 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.997281 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:41 crc kubenswrapper[4726]: I1001 16:00:41.997295 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:41Z","lastTransitionTime":"2025-10-01T16:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.100432 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.100520 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.100554 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.100587 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.100607 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:42Z","lastTransitionTime":"2025-10-01T16:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.203757 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.203826 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.203845 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.203869 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.203886 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:42Z","lastTransitionTime":"2025-10-01T16:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.307620 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.307712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.307730 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.307810 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.307829 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:42Z","lastTransitionTime":"2025-10-01T16:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.410185 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.410228 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.410242 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.410261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.410273 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:42Z","lastTransitionTime":"2025-10-01T16:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.513440 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.513485 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.513502 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.513524 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.513540 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:42Z","lastTransitionTime":"2025-10-01T16:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.615959 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.616017 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.616039 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.616104 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.616127 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:42Z","lastTransitionTime":"2025-10-01T16:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.719145 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.719277 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.719307 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.719341 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.719369 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:42Z","lastTransitionTime":"2025-10-01T16:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.807971 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.808020 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.808020 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:42 crc kubenswrapper[4726]: E1001 16:00:42.808147 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:42 crc kubenswrapper[4726]: E1001 16:00:42.808233 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:42 crc kubenswrapper[4726]: E1001 16:00:42.808329 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.824341 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.824400 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.824413 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.824435 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.824448 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:42Z","lastTransitionTime":"2025-10-01T16:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.931219 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.931291 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.931305 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.931325 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:42 crc kubenswrapper[4726]: I1001 16:00:42.931340 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:42Z","lastTransitionTime":"2025-10-01T16:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.004498 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.004537 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.004551 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.004567 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.004579 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:43Z","lastTransitionTime":"2025-10-01T16:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:43 crc kubenswrapper[4726]: E1001 16:00:43.028284 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:43Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.032816 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.032858 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.032875 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.032896 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.032913 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:43Z","lastTransitionTime":"2025-10-01T16:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:43 crc kubenswrapper[4726]: E1001 16:00:43.051328 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:43Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.055523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.055575 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.055593 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.055614 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.055632 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:43Z","lastTransitionTime":"2025-10-01T16:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:43 crc kubenswrapper[4726]: E1001 16:00:43.074088 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:43Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.077543 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.077584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.077600 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.077618 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.077629 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:43Z","lastTransitionTime":"2025-10-01T16:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:43 crc kubenswrapper[4726]: E1001 16:00:43.092634 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:43Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.096751 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.096793 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.096801 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.096816 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.096826 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:43Z","lastTransitionTime":"2025-10-01T16:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:43 crc kubenswrapper[4726]: E1001 16:00:43.112929 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"013da104-12c7-4736-af77-a6eb329b09c9\\\",\\\"systemUUID\\\":\\\"fd1dce9d-91f1-4195-90c9-2a6f634e684c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:43Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:43 crc kubenswrapper[4726]: E1001 16:00:43.113037 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.114883 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.114910 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.114919 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.114931 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.114940 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:43Z","lastTransitionTime":"2025-10-01T16:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.217635 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.217695 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.217711 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.217738 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.217755 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:43Z","lastTransitionTime":"2025-10-01T16:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.320729 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.320772 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.320784 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.320803 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.320814 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:43Z","lastTransitionTime":"2025-10-01T16:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.424320 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.424392 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.424418 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.424449 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.424472 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:43Z","lastTransitionTime":"2025-10-01T16:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.527836 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.527899 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.527916 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.527940 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.527957 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:43Z","lastTransitionTime":"2025-10-01T16:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.631170 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.631225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.631237 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.631254 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.631267 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:43Z","lastTransitionTime":"2025-10-01T16:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.733460 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.733506 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.733515 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.733535 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.733544 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:43Z","lastTransitionTime":"2025-10-01T16:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.807345 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:43 crc kubenswrapper[4726]: E1001 16:00:43.807487 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.836638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.836765 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.836783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.836810 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.836826 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:43Z","lastTransitionTime":"2025-10-01T16:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.939724 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.939802 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.939819 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.939844 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:43 crc kubenswrapper[4726]: I1001 16:00:43.939863 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:43Z","lastTransitionTime":"2025-10-01T16:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.043169 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.043249 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.043283 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.043335 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.043359 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:44Z","lastTransitionTime":"2025-10-01T16:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.146476 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.146540 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.146562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.146589 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.146606 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:44Z","lastTransitionTime":"2025-10-01T16:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.249219 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.249297 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.249317 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.249341 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.249360 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:44Z","lastTransitionTime":"2025-10-01T16:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.357502 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.358824 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.358880 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.358917 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.359027 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:44Z","lastTransitionTime":"2025-10-01T16:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.462109 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.462191 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.462210 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.462237 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.462259 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:44Z","lastTransitionTime":"2025-10-01T16:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.564669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.564766 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.564784 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.564812 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.564830 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:44Z","lastTransitionTime":"2025-10-01T16:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.667686 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.667785 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.667803 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.667827 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.667844 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:44Z","lastTransitionTime":"2025-10-01T16:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.770501 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.770570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.770611 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.770640 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.770660 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:44Z","lastTransitionTime":"2025-10-01T16:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.807738 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.807817 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:44 crc kubenswrapper[4726]: E1001 16:00:44.807908 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:44 crc kubenswrapper[4726]: E1001 16:00:44.807997 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.807741 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:44 crc kubenswrapper[4726]: E1001 16:00:44.808172 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.874524 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.874601 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.874625 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.874658 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.874680 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:44Z","lastTransitionTime":"2025-10-01T16:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.978383 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.978443 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.978462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.978486 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:44 crc kubenswrapper[4726]: I1001 16:00:44.978502 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:44Z","lastTransitionTime":"2025-10-01T16:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.081399 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.081508 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.081528 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.081552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.081568 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:45Z","lastTransitionTime":"2025-10-01T16:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.184267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.184303 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.184314 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.184329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.184341 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:45Z","lastTransitionTime":"2025-10-01T16:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.287380 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.287425 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.287436 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.287452 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.287462 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:45Z","lastTransitionTime":"2025-10-01T16:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.390564 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.390622 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.390639 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.390663 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.390701 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:45Z","lastTransitionTime":"2025-10-01T16:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.493628 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.493696 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.493719 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.493747 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.493769 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:45Z","lastTransitionTime":"2025-10-01T16:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.596566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.596644 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.596665 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.596692 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.596712 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:45Z","lastTransitionTime":"2025-10-01T16:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.699204 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.699285 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.699304 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.699326 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.699342 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:45Z","lastTransitionTime":"2025-10-01T16:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.801867 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.801907 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.801915 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.801928 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.801937 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:45Z","lastTransitionTime":"2025-10-01T16:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.807685 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:45 crc kubenswrapper[4726]: E1001 16:00:45.807861 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.904647 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.904719 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.904743 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.904781 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:45 crc kubenswrapper[4726]: I1001 16:00:45.904807 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:45Z","lastTransitionTime":"2025-10-01T16:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.007611 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.007664 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.007682 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.007707 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.007731 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:46Z","lastTransitionTime":"2025-10-01T16:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.110536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.110583 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.110598 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.110617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.110648 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:46Z","lastTransitionTime":"2025-10-01T16:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.213323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.213391 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.213404 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.213421 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.213432 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:46Z","lastTransitionTime":"2025-10-01T16:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.316216 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.316285 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.316302 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.316323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.316338 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:46Z","lastTransitionTime":"2025-10-01T16:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.419798 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.419858 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.419877 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.419903 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.419923 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:46Z","lastTransitionTime":"2025-10-01T16:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.522342 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.522382 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.522419 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.522439 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.522454 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:46Z","lastTransitionTime":"2025-10-01T16:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.625591 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.625640 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.625674 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.625692 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.625703 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:46Z","lastTransitionTime":"2025-10-01T16:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.728278 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.728315 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.728327 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.728343 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.728355 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:46Z","lastTransitionTime":"2025-10-01T16:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.807721 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:46 crc kubenswrapper[4726]: E1001 16:00:46.807862 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.807735 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.807902 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:46 crc kubenswrapper[4726]: E1001 16:00:46.807951 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:46 crc kubenswrapper[4726]: E1001 16:00:46.808386 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.831153 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.831257 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.831276 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.831303 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.831323 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:46Z","lastTransitionTime":"2025-10-01T16:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.934475 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.934863 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.935006 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.935158 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:46 crc kubenswrapper[4726]: I1001 16:00:46.935287 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:46Z","lastTransitionTime":"2025-10-01T16:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.038444 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.038482 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.038492 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.038506 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.038516 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:47Z","lastTransitionTime":"2025-10-01T16:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.141331 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.141869 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.142104 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.142341 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.142505 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:47Z","lastTransitionTime":"2025-10-01T16:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.245845 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.246151 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.246239 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.246307 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.246397 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:47Z","lastTransitionTime":"2025-10-01T16:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.348634 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.348686 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.348700 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.348718 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.348731 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:47Z","lastTransitionTime":"2025-10-01T16:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.450773 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.450835 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.450844 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.450858 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.450867 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:47Z","lastTransitionTime":"2025-10-01T16:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.554104 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.554209 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.554229 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.554294 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.554314 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:47Z","lastTransitionTime":"2025-10-01T16:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.657434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.657495 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.657527 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.657557 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.657577 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:47Z","lastTransitionTime":"2025-10-01T16:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.761167 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.761241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.761276 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.761307 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.761326 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:47Z","lastTransitionTime":"2025-10-01T16:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.807993 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:47 crc kubenswrapper[4726]: E1001 16:00:47.808299 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.831410 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:47Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.849630 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d1252d7e372c8cd41e8f9d45ac6f818344ee02850f40fbc1b0d3c58b99af386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:47Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.864379 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.864631 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.864772 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.864925 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.865094 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:47Z","lastTransitionTime":"2025-10-01T16:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.875966 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cx25v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21c32caa-f2fd-4c2a-a766-52f3fc81a0ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c52d47679be14f1d39ecbae1c0bf53edefefd8515bbd6324d711a81692a611c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0615af8ee3ce77457613c27f0844f959ee262956558e704d8c5091bd593ccd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6765b3ebdca911b288c863f2b37e65137461b2a71b7382e4523d2f53fcdc211d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2818e5c108d928bad5676ab56b63eaf16f993e16f1f3b13a52d1914bc91adfbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://384d894d7e8efe72fd523fa96def224d39807d1738a2972c5b7f6d1cbc577672\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73f21620749e2305324956c90a32f0acc0a37ab1ba906252a210fea5569e9f54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4179ab9d712fcd36851fac1c2172e55d6d04b472a20b81ec85a9ce04159ba710\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkm7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cx25v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:47Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.893913 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a22e7a8-726a-4538-a27d-44c58f89f0b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gccgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4pzd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:47Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.930492 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee1e9645-377a-4bb2-b2ea-da72e745ec47\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48cf222d354b944a81a8f9368bca357c490b4cccd218a77c7a6b2edba3923d37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5c9d8c826f12a7c2e9807bedf5cdc779cc909e3aeae02fd84b3ef9803b19bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://123c98db1c88c13a245bf8351f4b774360c7da9bc25a4af6555dec68896f88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8953b3e39038c3f4d5e83cff594bc7003f2881938115d0234c0b2c53cefb210\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15fb7ef5d7d02b41f93de3d9cd0daade9f34b3ca164fa949d8727dfe253e7661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbc61d45e451257595e0768e02c17ac376197b25423eaaadc0c256a395ea578\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbc61d45e451257595e0768e02c17ac376197b25423eaaadc0c256a395ea578\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c95e8700c1fc896258d1d15a49ff6364853ff7ef50445c0ba5778efd16c86fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c95e8700c1fc896258d1d15a49ff6364853ff7ef50445c0ba5778efd16c86fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c8ed80846072f8edfd40226b620e585fa8b8e2f131424b30be12d0e7d27f7c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ed80846072f8edfd40226b620e585fa8b8e2f131424b30be12d0e7d27f7c68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:47Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.954657 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3f39d36-80fe-408a-ba1d-091859f541e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://572b3d0e5e5e85de861e0ab929a99e8f8fc9b4800dcf4b97996e7537dae08fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f63cafe35315430bba7ec56c236c66536bcf4bc6f36864e99556c85a2507c98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a9b3b970b4dcf923212eeb5f1dd30ac707b0131749e58a0b40f08d40c9ad35b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d19fac6fc428e65d71674f0d956735572615dd7bb3707021e2e3b2f46c467187\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:47Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.968539 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.968612 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.968636 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.968667 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.968691 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:47Z","lastTransitionTime":"2025-10-01T16:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.974641 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af78c660dbbdfe1c72e4155497419f40525afe79ef152400b4f625742eefee0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef29ab20892794e73aecb8bb7f6d59fe51b702cb133b097ec2f3c6afbec5d8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:47Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:47 crc kubenswrapper[4726]: I1001 16:00:47.994471 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:47Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.010368 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5sdsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e777c1f8-99c8-4dd9-a920-7bcfcdd5e229\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7f0715b0e20811ec82e6fdf6b42adb4d7a216f9217cda2ae68a48fe459a2e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5v77c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5sdsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:48Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.022293 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fmbgk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf1db398-f129-4277-8527-e9117daec3a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11cc45cae3594cc354ead34f62647761626e4ea85d2aa783c264798dd9655eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lgm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fmbgk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:48Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.034836 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8be9689-98f6-45f3-9523-3b66f3a55c1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc7645d98b587e9a0822e28fd84cf6e3807aefd79419ce0f30768a86b8444da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://817d0112385e1a139ad40977583129512dc9f19e6cd39af2e3021a72966192a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwhdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vdljl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:48Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.057969 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h8cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9257aec-e319-4261-ae4c-dec88468b680\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e127584cf73c7f9ea20da97dce5f50fb86b8f9fde6ae383172834d425bb1e033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:18Z\\\",\\\"message\\\":\\\"2025-10-01T15:59:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14\\\\n2025-10-01T15:59:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7202bd09-63a3-4d3a-a372-b7e284242a14 to /host/opt/cni/bin/\\\\n2025-10-01T15:59:33Z [verbose] multus-daemon started\\\\n2025-10-01T15:59:33Z [verbose] Readiness Indicator file check\\\\n2025-10-01T16:00:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T16:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67cvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h8cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:48Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.071138 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.071185 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.071199 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.071221 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.071235 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:48Z","lastTransitionTime":"2025-10-01T16:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.077140 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00d8cf08-6eff-420a-be98-e0dd45908be6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee0d4a9e1b0138a84c2bc714493d778829840a609b535e2274c1f50cca33efa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jllb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x5gdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:48Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.110023 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T16:00:34Z\\\",\\\"message\\\":\\\"ssqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 16:00:34.715200 6847 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 16:00:34.715257 6847 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 16:00:34.715288 6847 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 16:00:34.715698 6847 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 16:00:34.715725 6847 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 16:00:34.715741 6847 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 16:00:34.715746 6847 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 16:00:34.715818 6847 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 16:00:34.715835 6847 factory.go:656] Stopping watch factory\\\\nI1001 16:00:34.715835 6847 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 16:00:34.715846 6847 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 16:00:34.715859 6847 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T16:00:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znnhm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zp4jh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:48Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.125574 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3064c56-dd95-414e-83a1-8876f73866fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4179687cd0b6fa600a3efc7b5f5761a15fb6df23af0f1b0c94170bb106b3154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8708d6c7f25de9cc09ddce04f12fcf28f5e13b3b0e0418e5a4852c0d58bfcb56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:48Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.140740 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fcb45bd-ff17-49be-927e-0406e37bb00d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20aecbc9dc23270bbc0b50f9974f67caea5e9475f721db325ce44fb39046fc9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10616ac42ca73899e22bb71042cf608f01f8415e4432a00fcf0e1321d0f5b660\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d3a9e2463f96116d2482f06e35dcf379bb1db90542d8b13f4ace78bc4533a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e823831da470977e38878df3bf428082392c7f6c49edc2045211afdf269906bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bbd05ccc6b56f829239325c31d3e32ea5650c2938401c8e910b43cfdf53425e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"ey\\\\\\\"\\\\nI1001 15:59:29.435264 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:59:29.477103 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:59:29.477147 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:59:29.477194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:59:29.477209 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:59:29.495203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:59:29.495242 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:59:29.495273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:59:29.495280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:59:29.495287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:59:29.495293 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:59:29.495282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:59:29.498918 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:59:29.500268 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2868296734/tls.crt::/tmp/serving-cert-2868296734/tls.key\\\\\\\"\\\\nI1001 15:59:29.500299 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 15:59:29.500329 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://970927fdde0a917e717a62074f354766ba94d99476b0c3402aff2dfddcf3fff2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2789859596c91f91c5ce7b76d6f20c8e3ef24d7c3c9eb86ea431018746cfcca4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:48Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.153998 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12eb575b-72e7-4bb2-b9de-ce924d47db1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T16:00:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9fe86b3d6a1acd9512663da493ea130d4112b79d2ea24e5d94eeb76e81d852\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f94a8081041e5968c831dce9988a771386310ae721226c4f58baa6a2c1af47b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565eeb4bdc62a6f23f5a056700c8677e7de8bac37bd7cc766d4b8637c7fe3f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4236997ad0e6b4580cea828c731d2e0118650aa145ead54e34c72554cbc899e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:59:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:59:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:59:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:48Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.165797 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a23eb5801e6170a8bd75e14c82adaf25dde80934c424f8532acca73cde19e8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:59:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:48Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.174020 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.174102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.174116 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.174132 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.174142 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:48Z","lastTransitionTime":"2025-10-01T16:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.179210 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:59:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T16:00:48Z is after 2025-08-24T17:21:41Z" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.277426 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.277501 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.277517 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.277539 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.277556 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:48Z","lastTransitionTime":"2025-10-01T16:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.379736 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.379791 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.379801 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.379813 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.379822 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:48Z","lastTransitionTime":"2025-10-01T16:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.482988 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.483096 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.483125 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.483144 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.483193 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:48Z","lastTransitionTime":"2025-10-01T16:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.586510 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.586664 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.586697 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.586772 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.586800 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:48Z","lastTransitionTime":"2025-10-01T16:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.689128 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.689213 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.689253 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.689294 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.689319 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:48Z","lastTransitionTime":"2025-10-01T16:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.792729 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.792802 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.792820 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.792843 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.792860 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:48Z","lastTransitionTime":"2025-10-01T16:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.807393 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.807447 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.807402 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:48 crc kubenswrapper[4726]: E1001 16:00:48.807574 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:48 crc kubenswrapper[4726]: E1001 16:00:48.807658 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:48 crc kubenswrapper[4726]: E1001 16:00:48.807729 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.896476 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.896532 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.896550 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.896573 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.896590 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:48Z","lastTransitionTime":"2025-10-01T16:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.999176 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.999277 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.999296 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.999318 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:48 crc kubenswrapper[4726]: I1001 16:00:48.999335 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:48Z","lastTransitionTime":"2025-10-01T16:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.101441 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.101504 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.101523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.101546 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.101562 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:49Z","lastTransitionTime":"2025-10-01T16:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.204409 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.204467 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.204478 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.204496 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.204509 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:49Z","lastTransitionTime":"2025-10-01T16:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.307725 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.307791 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.307804 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.307820 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.307853 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:49Z","lastTransitionTime":"2025-10-01T16:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.411302 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.411364 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.411375 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.411392 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.411402 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:49Z","lastTransitionTime":"2025-10-01T16:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.442402 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs\") pod \"network-metrics-daemon-4pzd8\" (UID: \"2a22e7a8-726a-4538-a27d-44c58f89f0b1\") " pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:49 crc kubenswrapper[4726]: E1001 16:00:49.442596 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 16:00:49 crc kubenswrapper[4726]: E1001 16:00:49.442713 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs podName:2a22e7a8-726a-4538-a27d-44c58f89f0b1 nodeName:}" failed. No retries permitted until 2025-10-01 16:01:53.442684018 +0000 UTC m=+166.344236625 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs") pod "network-metrics-daemon-4pzd8" (UID: "2a22e7a8-726a-4538-a27d-44c58f89f0b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.552590 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.552645 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.552668 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.552694 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.552716 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:49Z","lastTransitionTime":"2025-10-01T16:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.656496 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.656543 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.656551 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.656566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.656575 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:49Z","lastTransitionTime":"2025-10-01T16:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.759457 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.759529 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.759567 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.759603 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.759626 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:49Z","lastTransitionTime":"2025-10-01T16:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.807938 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:49 crc kubenswrapper[4726]: E1001 16:00:49.808588 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.808962 4726 scope.go:117] "RemoveContainer" containerID="04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be" Oct 01 16:00:49 crc kubenswrapper[4726]: E1001 16:00:49.809252 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.863032 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.863133 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.863152 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.863176 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.863194 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:49Z","lastTransitionTime":"2025-10-01T16:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.965337 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.965377 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.965386 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.965399 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:49 crc kubenswrapper[4726]: I1001 16:00:49.965410 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:49Z","lastTransitionTime":"2025-10-01T16:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.068530 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.068592 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.068610 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.068635 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.068654 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:50Z","lastTransitionTime":"2025-10-01T16:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.171993 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.172083 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.172102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.172126 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.172142 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:50Z","lastTransitionTime":"2025-10-01T16:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.275669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.275721 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.275743 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.275774 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.275800 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:50Z","lastTransitionTime":"2025-10-01T16:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.377822 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.377865 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.377876 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.377892 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.377903 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:50Z","lastTransitionTime":"2025-10-01T16:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.480979 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.481045 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.481160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.481184 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.481202 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:50Z","lastTransitionTime":"2025-10-01T16:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.584290 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.584380 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.584400 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.584422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.584439 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:50Z","lastTransitionTime":"2025-10-01T16:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.687701 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.687753 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.687776 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.687805 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.687828 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:50Z","lastTransitionTime":"2025-10-01T16:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.790468 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.790506 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.790517 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.790532 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.790543 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:50Z","lastTransitionTime":"2025-10-01T16:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.807635 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.807705 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.807867 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:50 crc kubenswrapper[4726]: E1001 16:00:50.808241 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:50 crc kubenswrapper[4726]: E1001 16:00:50.808378 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:50 crc kubenswrapper[4726]: E1001 16:00:50.809510 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.892925 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.893249 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.893337 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.893422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.893517 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:50Z","lastTransitionTime":"2025-10-01T16:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.996241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.996361 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.996421 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.996452 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:50 crc kubenswrapper[4726]: I1001 16:00:50.996470 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:50Z","lastTransitionTime":"2025-10-01T16:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.098898 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.098956 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.098974 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.098999 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.099018 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:51Z","lastTransitionTime":"2025-10-01T16:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.202170 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.202236 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.202245 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.202258 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.202266 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:51Z","lastTransitionTime":"2025-10-01T16:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.304254 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.304314 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.304324 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.304341 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.304354 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:51Z","lastTransitionTime":"2025-10-01T16:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.407330 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.407375 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.407393 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.407415 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.407432 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:51Z","lastTransitionTime":"2025-10-01T16:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.509857 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.509930 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.509947 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.509975 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.509993 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:51Z","lastTransitionTime":"2025-10-01T16:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.613249 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.613315 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.613332 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.613354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.613372 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:51Z","lastTransitionTime":"2025-10-01T16:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.716328 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.716429 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.716493 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.716525 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.716547 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:51Z","lastTransitionTime":"2025-10-01T16:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.807182 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:51 crc kubenswrapper[4726]: E1001 16:00:51.807351 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.819542 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.819615 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.819634 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.819659 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.819678 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:51Z","lastTransitionTime":"2025-10-01T16:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.922721 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.922762 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.922774 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.922789 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:51 crc kubenswrapper[4726]: I1001 16:00:51.922801 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:51Z","lastTransitionTime":"2025-10-01T16:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.025885 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.025958 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.025980 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.026008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.026028 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:52Z","lastTransitionTime":"2025-10-01T16:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.129287 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.129332 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.129344 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.129360 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.129374 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:52Z","lastTransitionTime":"2025-10-01T16:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.232552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.232624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.232645 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.232671 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.232689 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:52Z","lastTransitionTime":"2025-10-01T16:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.336221 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.336272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.336294 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.336321 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.336367 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:52Z","lastTransitionTime":"2025-10-01T16:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.444243 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.444310 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.444339 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.444370 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.444389 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:52Z","lastTransitionTime":"2025-10-01T16:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.547807 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.547872 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.547887 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.547918 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.547938 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:52Z","lastTransitionTime":"2025-10-01T16:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.651233 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.651307 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.651329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.651357 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.651377 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:52Z","lastTransitionTime":"2025-10-01T16:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.754194 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.754261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.754278 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.754302 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.754320 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:52Z","lastTransitionTime":"2025-10-01T16:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.807361 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.807387 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.807530 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:52 crc kubenswrapper[4726]: E1001 16:00:52.807697 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:52 crc kubenswrapper[4726]: E1001 16:00:52.807860 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:52 crc kubenswrapper[4726]: E1001 16:00:52.807959 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.856986 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.857150 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.857190 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.857223 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.857244 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:52Z","lastTransitionTime":"2025-10-01T16:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.960522 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.960581 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.960600 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.960622 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:52 crc kubenswrapper[4726]: I1001 16:00:52.960638 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:52Z","lastTransitionTime":"2025-10-01T16:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.063720 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.063814 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.063838 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.063867 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.063888 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:53Z","lastTransitionTime":"2025-10-01T16:00:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.167296 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.167370 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.167395 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.167429 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.167453 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:53Z","lastTransitionTime":"2025-10-01T16:00:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.239025 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.239146 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.239197 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.239231 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.239257 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T16:00:53Z","lastTransitionTime":"2025-10-01T16:00:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.310722 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c"] Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.311152 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.313943 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.314426 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.314769 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.315789 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.354983 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=84.354958659 podStartE2EDuration="1m24.354958659s" podCreationTimestamp="2025-10-01 15:59:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:00:53.332805529 +0000 UTC m=+106.234358106" watchObservedRunningTime="2025-10-01 16:00:53.354958659 +0000 UTC m=+106.256511246" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.369123 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=50.369101455 podStartE2EDuration="50.369101455s" podCreationTimestamp="2025-10-01 16:00:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:00:53.354942068 +0000 UTC m=+106.256494695" watchObservedRunningTime="2025-10-01 16:00:53.369101455 +0000 UTC m=+106.270654082" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.403555 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-h8cql" podStartSLOduration=83.403534991 podStartE2EDuration="1m23.403534991s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:00:53.403241963 +0000 UTC m=+106.304794550" watchObservedRunningTime="2025-10-01 16:00:53.403534991 +0000 UTC m=+106.305087578" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.480010 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podStartSLOduration=83.479991126 podStartE2EDuration="1m23.479991126s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:00:53.419333142 +0000 UTC m=+106.320885739" watchObservedRunningTime="2025-10-01 16:00:53.479991126 +0000 UTC m=+106.381543703" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.491815 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=35.49179555 podStartE2EDuration="35.49179555s" podCreationTimestamp="2025-10-01 16:00:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:00:53.491205344 +0000 UTC m=+106.392757921" watchObservedRunningTime="2025-10-01 16:00:53.49179555 +0000 UTC m=+106.393348127" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.492662 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.492695 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.492710 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.492735 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.492752 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.540102 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-cx25v" podStartSLOduration=83.540083435 podStartE2EDuration="1m23.540083435s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:00:53.529857792 +0000 UTC m=+106.431410369" watchObservedRunningTime="2025-10-01 16:00:53.540083435 +0000 UTC m=+106.441636002" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.576015 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=84.57599758 podStartE2EDuration="1m24.57599758s" podCreationTimestamp="2025-10-01 15:59:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:00:53.575541298 +0000 UTC m=+106.477093885" watchObservedRunningTime="2025-10-01 16:00:53.57599758 +0000 UTC m=+106.477550157" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.593314 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.593620 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.593765 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.594033 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.594213 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.594166 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.594304 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.594952 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.601772 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.617211 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-5sdsp" podStartSLOduration=83.617185956 podStartE2EDuration="1m23.617185956s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:00:53.616865538 +0000 UTC m=+106.518418125" watchObservedRunningTime="2025-10-01 16:00:53.617185956 +0000 UTC m=+106.518738533" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.618467 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6c1e36a-aeeb-471d-b335-267efd3d3ff2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-m526c\" (UID: \"a6c1e36a-aeeb-471d-b335-267efd3d3ff2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.627956 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.662691 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=12.662676667 podStartE2EDuration="12.662676667s" podCreationTimestamp="2025-10-01 16:00:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:00:53.661274659 +0000 UTC m=+106.562827236" watchObservedRunningTime="2025-10-01 16:00:53.662676667 +0000 UTC m=+106.564229244" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.689061 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-fmbgk" podStartSLOduration=83.689027968 podStartE2EDuration="1m23.689027968s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:00:53.688819992 +0000 UTC m=+106.590372579" watchObservedRunningTime="2025-10-01 16:00:53.689027968 +0000 UTC m=+106.590580545" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.689161 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vdljl" podStartSLOduration=83.689155361 podStartE2EDuration="1m23.689155361s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:00:53.67710005 +0000 UTC m=+106.578652637" watchObservedRunningTime="2025-10-01 16:00:53.689155361 +0000 UTC m=+106.590707938" Oct 01 16:00:53 crc kubenswrapper[4726]: I1001 16:00:53.807857 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:53 crc kubenswrapper[4726]: E1001 16:00:53.807993 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:54 crc kubenswrapper[4726]: I1001 16:00:54.394989 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" event={"ID":"a6c1e36a-aeeb-471d-b335-267efd3d3ff2","Type":"ContainerStarted","Data":"ae1176a8b4472a5acb1b818f4e4ff55a01a176311390d1b5c3411257878b77dc"} Oct 01 16:00:54 crc kubenswrapper[4726]: I1001 16:00:54.395069 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" event={"ID":"a6c1e36a-aeeb-471d-b335-267efd3d3ff2","Type":"ContainerStarted","Data":"8ad62e53f7d9aac60e2eb63d0b86631ca1f6c49d3a45bcfe6abdb6d8671809d1"} Oct 01 16:00:54 crc kubenswrapper[4726]: I1001 16:00:54.417791 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m526c" podStartSLOduration=84.417768719 podStartE2EDuration="1m24.417768719s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:00:54.41669405 +0000 UTC m=+107.318246627" watchObservedRunningTime="2025-10-01 16:00:54.417768719 +0000 UTC m=+107.319321306" Oct 01 16:00:54 crc kubenswrapper[4726]: I1001 16:00:54.806961 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:54 crc kubenswrapper[4726]: I1001 16:00:54.807061 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:54 crc kubenswrapper[4726]: E1001 16:00:54.807110 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:54 crc kubenswrapper[4726]: I1001 16:00:54.807129 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:54 crc kubenswrapper[4726]: E1001 16:00:54.807213 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:54 crc kubenswrapper[4726]: E1001 16:00:54.807346 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:55 crc kubenswrapper[4726]: I1001 16:00:55.807429 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:55 crc kubenswrapper[4726]: E1001 16:00:55.807677 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:56 crc kubenswrapper[4726]: I1001 16:00:56.807686 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:56 crc kubenswrapper[4726]: I1001 16:00:56.807749 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:56 crc kubenswrapper[4726]: I1001 16:00:56.807821 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:56 crc kubenswrapper[4726]: E1001 16:00:56.807997 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:56 crc kubenswrapper[4726]: E1001 16:00:56.808232 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:56 crc kubenswrapper[4726]: E1001 16:00:56.808328 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:57 crc kubenswrapper[4726]: I1001 16:00:57.807831 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:57 crc kubenswrapper[4726]: E1001 16:00:57.809680 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:00:58 crc kubenswrapper[4726]: I1001 16:00:58.807836 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:00:58 crc kubenswrapper[4726]: I1001 16:00:58.808096 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:00:58 crc kubenswrapper[4726]: E1001 16:00:58.808413 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:00:58 crc kubenswrapper[4726]: I1001 16:00:58.808106 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:00:58 crc kubenswrapper[4726]: E1001 16:00:58.808588 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:00:58 crc kubenswrapper[4726]: E1001 16:00:58.808780 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:00:59 crc kubenswrapper[4726]: I1001 16:00:59.807442 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:00:59 crc kubenswrapper[4726]: E1001 16:00:59.807689 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:00 crc kubenswrapper[4726]: I1001 16:01:00.807034 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:00 crc kubenswrapper[4726]: I1001 16:01:00.807163 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:00 crc kubenswrapper[4726]: I1001 16:01:00.807400 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:00 crc kubenswrapper[4726]: E1001 16:01:00.807695 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:00 crc kubenswrapper[4726]: E1001 16:01:00.807936 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:00 crc kubenswrapper[4726]: E1001 16:01:00.808028 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:01 crc kubenswrapper[4726]: I1001 16:01:01.807698 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:01 crc kubenswrapper[4726]: E1001 16:01:01.807946 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:02 crc kubenswrapper[4726]: I1001 16:01:02.807018 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:02 crc kubenswrapper[4726]: I1001 16:01:02.807465 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:02 crc kubenswrapper[4726]: I1001 16:01:02.807494 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:02 crc kubenswrapper[4726]: E1001 16:01:02.807763 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:02 crc kubenswrapper[4726]: I1001 16:01:02.807791 4726 scope.go:117] "RemoveContainer" containerID="04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be" Oct 01 16:01:02 crc kubenswrapper[4726]: E1001 16:01:02.807900 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:02 crc kubenswrapper[4726]: E1001 16:01:02.807985 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:02 crc kubenswrapper[4726]: E1001 16:01:02.808216 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" Oct 01 16:01:03 crc kubenswrapper[4726]: I1001 16:01:03.808132 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:03 crc kubenswrapper[4726]: E1001 16:01:03.808293 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:04 crc kubenswrapper[4726]: I1001 16:01:04.807335 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:04 crc kubenswrapper[4726]: I1001 16:01:04.807315 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:04 crc kubenswrapper[4726]: I1001 16:01:04.807446 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:04 crc kubenswrapper[4726]: E1001 16:01:04.807547 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:04 crc kubenswrapper[4726]: E1001 16:01:04.807945 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:04 crc kubenswrapper[4726]: E1001 16:01:04.808305 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:05 crc kubenswrapper[4726]: I1001 16:01:05.433848 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h8cql_e9257aec-e319-4261-ae4c-dec88468b680/kube-multus/1.log" Oct 01 16:01:05 crc kubenswrapper[4726]: I1001 16:01:05.434969 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h8cql_e9257aec-e319-4261-ae4c-dec88468b680/kube-multus/0.log" Oct 01 16:01:05 crc kubenswrapper[4726]: I1001 16:01:05.435092 4726 generic.go:334] "Generic (PLEG): container finished" podID="e9257aec-e319-4261-ae4c-dec88468b680" containerID="e127584cf73c7f9ea20da97dce5f50fb86b8f9fde6ae383172834d425bb1e033" exitCode=1 Oct 01 16:01:05 crc kubenswrapper[4726]: I1001 16:01:05.435171 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h8cql" event={"ID":"e9257aec-e319-4261-ae4c-dec88468b680","Type":"ContainerDied","Data":"e127584cf73c7f9ea20da97dce5f50fb86b8f9fde6ae383172834d425bb1e033"} Oct 01 16:01:05 crc kubenswrapper[4726]: I1001 16:01:05.435275 4726 scope.go:117] "RemoveContainer" containerID="e6e056087d0618fe3c357858c81ec45c28f8cceeb736c0db6c18771e12084ead" Oct 01 16:01:05 crc kubenswrapper[4726]: I1001 16:01:05.436257 4726 scope.go:117] "RemoveContainer" containerID="e127584cf73c7f9ea20da97dce5f50fb86b8f9fde6ae383172834d425bb1e033" Oct 01 16:01:05 crc kubenswrapper[4726]: E1001 16:01:05.436594 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-h8cql_openshift-multus(e9257aec-e319-4261-ae4c-dec88468b680)\"" pod="openshift-multus/multus-h8cql" podUID="e9257aec-e319-4261-ae4c-dec88468b680" Oct 01 16:01:05 crc kubenswrapper[4726]: I1001 16:01:05.807481 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:05 crc kubenswrapper[4726]: E1001 16:01:05.807640 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:06 crc kubenswrapper[4726]: I1001 16:01:06.440724 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h8cql_e9257aec-e319-4261-ae4c-dec88468b680/kube-multus/1.log" Oct 01 16:01:06 crc kubenswrapper[4726]: I1001 16:01:06.807924 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:06 crc kubenswrapper[4726]: E1001 16:01:06.808068 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:06 crc kubenswrapper[4726]: I1001 16:01:06.807947 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:06 crc kubenswrapper[4726]: I1001 16:01:06.807928 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:06 crc kubenswrapper[4726]: E1001 16:01:06.808394 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:06 crc kubenswrapper[4726]: E1001 16:01:06.808494 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:07 crc kubenswrapper[4726]: E1001 16:01:07.798109 4726 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 01 16:01:07 crc kubenswrapper[4726]: I1001 16:01:07.807677 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:07 crc kubenswrapper[4726]: E1001 16:01:07.809413 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:07 crc kubenswrapper[4726]: E1001 16:01:07.924264 4726 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 16:01:08 crc kubenswrapper[4726]: I1001 16:01:08.807795 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:08 crc kubenswrapper[4726]: E1001 16:01:08.807939 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:08 crc kubenswrapper[4726]: I1001 16:01:08.807829 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:08 crc kubenswrapper[4726]: I1001 16:01:08.807802 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:08 crc kubenswrapper[4726]: E1001 16:01:08.808030 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:08 crc kubenswrapper[4726]: E1001 16:01:08.808241 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:09 crc kubenswrapper[4726]: I1001 16:01:09.807348 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:09 crc kubenswrapper[4726]: E1001 16:01:09.807588 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:10 crc kubenswrapper[4726]: I1001 16:01:10.807141 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:10 crc kubenswrapper[4726]: I1001 16:01:10.807199 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:10 crc kubenswrapper[4726]: I1001 16:01:10.807276 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:10 crc kubenswrapper[4726]: E1001 16:01:10.807325 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:10 crc kubenswrapper[4726]: E1001 16:01:10.807498 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:10 crc kubenswrapper[4726]: E1001 16:01:10.807728 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:11 crc kubenswrapper[4726]: I1001 16:01:11.807309 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:11 crc kubenswrapper[4726]: E1001 16:01:11.807810 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:12 crc kubenswrapper[4726]: I1001 16:01:12.807468 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:12 crc kubenswrapper[4726]: I1001 16:01:12.807558 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:12 crc kubenswrapper[4726]: E1001 16:01:12.808042 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:12 crc kubenswrapper[4726]: I1001 16:01:12.807613 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:12 crc kubenswrapper[4726]: E1001 16:01:12.808323 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:12 crc kubenswrapper[4726]: E1001 16:01:12.808439 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:12 crc kubenswrapper[4726]: E1001 16:01:12.926115 4726 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 16:01:13 crc kubenswrapper[4726]: I1001 16:01:13.808142 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:13 crc kubenswrapper[4726]: E1001 16:01:13.808524 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:13 crc kubenswrapper[4726]: I1001 16:01:13.809720 4726 scope.go:117] "RemoveContainer" containerID="04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be" Oct 01 16:01:13 crc kubenswrapper[4726]: E1001 16:01:13.810015 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zp4jh_openshift-ovn-kubernetes(48c2aeef-b0f7-41b1-8a03-519171bc0a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" Oct 01 16:01:14 crc kubenswrapper[4726]: I1001 16:01:14.808121 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:14 crc kubenswrapper[4726]: I1001 16:01:14.808122 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:14 crc kubenswrapper[4726]: I1001 16:01:14.808354 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:14 crc kubenswrapper[4726]: E1001 16:01:14.808502 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:14 crc kubenswrapper[4726]: E1001 16:01:14.808626 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:14 crc kubenswrapper[4726]: E1001 16:01:14.808819 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:15 crc kubenswrapper[4726]: I1001 16:01:15.807008 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:15 crc kubenswrapper[4726]: E1001 16:01:15.807263 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:15 crc kubenswrapper[4726]: I1001 16:01:15.807787 4726 scope.go:117] "RemoveContainer" containerID="e127584cf73c7f9ea20da97dce5f50fb86b8f9fde6ae383172834d425bb1e033" Oct 01 16:01:16 crc kubenswrapper[4726]: I1001 16:01:16.477552 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h8cql_e9257aec-e319-4261-ae4c-dec88468b680/kube-multus/1.log" Oct 01 16:01:16 crc kubenswrapper[4726]: I1001 16:01:16.477628 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h8cql" event={"ID":"e9257aec-e319-4261-ae4c-dec88468b680","Type":"ContainerStarted","Data":"c061620e1547381b1e62f4d7068ac10ea23bd2fb250a3ce42f6f14cd2e8c727b"} Oct 01 16:01:16 crc kubenswrapper[4726]: I1001 16:01:16.807639 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:16 crc kubenswrapper[4726]: I1001 16:01:16.807759 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:16 crc kubenswrapper[4726]: E1001 16:01:16.807767 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:16 crc kubenswrapper[4726]: E1001 16:01:16.807979 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:16 crc kubenswrapper[4726]: I1001 16:01:16.808024 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:16 crc kubenswrapper[4726]: E1001 16:01:16.808110 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:17 crc kubenswrapper[4726]: I1001 16:01:17.807280 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:17 crc kubenswrapper[4726]: E1001 16:01:17.809392 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:17 crc kubenswrapper[4726]: E1001 16:01:17.927662 4726 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 16:01:18 crc kubenswrapper[4726]: I1001 16:01:18.806967 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:18 crc kubenswrapper[4726]: I1001 16:01:18.807264 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:18 crc kubenswrapper[4726]: E1001 16:01:18.807294 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:18 crc kubenswrapper[4726]: E1001 16:01:18.807427 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:18 crc kubenswrapper[4726]: I1001 16:01:18.806967 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:18 crc kubenswrapper[4726]: E1001 16:01:18.807524 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:19 crc kubenswrapper[4726]: I1001 16:01:19.807392 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:19 crc kubenswrapper[4726]: E1001 16:01:19.807789 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:20 crc kubenswrapper[4726]: I1001 16:01:20.807538 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:20 crc kubenswrapper[4726]: I1001 16:01:20.807605 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:20 crc kubenswrapper[4726]: I1001 16:01:20.807545 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:20 crc kubenswrapper[4726]: E1001 16:01:20.807705 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:20 crc kubenswrapper[4726]: E1001 16:01:20.807814 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:20 crc kubenswrapper[4726]: E1001 16:01:20.807913 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:21 crc kubenswrapper[4726]: I1001 16:01:21.808015 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:21 crc kubenswrapper[4726]: E1001 16:01:21.808166 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:22 crc kubenswrapper[4726]: I1001 16:01:22.807345 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:22 crc kubenswrapper[4726]: I1001 16:01:22.807353 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:22 crc kubenswrapper[4726]: I1001 16:01:22.807496 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:22 crc kubenswrapper[4726]: E1001 16:01:22.807691 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:22 crc kubenswrapper[4726]: E1001 16:01:22.807786 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:22 crc kubenswrapper[4726]: E1001 16:01:22.807894 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:22 crc kubenswrapper[4726]: E1001 16:01:22.929559 4726 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 16:01:23 crc kubenswrapper[4726]: I1001 16:01:23.808121 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:23 crc kubenswrapper[4726]: E1001 16:01:23.808291 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:24 crc kubenswrapper[4726]: I1001 16:01:24.806998 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:24 crc kubenswrapper[4726]: I1001 16:01:24.807101 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:24 crc kubenswrapper[4726]: I1001 16:01:24.807118 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:24 crc kubenswrapper[4726]: E1001 16:01:24.807236 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:24 crc kubenswrapper[4726]: E1001 16:01:24.807303 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:24 crc kubenswrapper[4726]: E1001 16:01:24.807475 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:25 crc kubenswrapper[4726]: I1001 16:01:25.808037 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:25 crc kubenswrapper[4726]: E1001 16:01:25.808303 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:26 crc kubenswrapper[4726]: I1001 16:01:26.807480 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:26 crc kubenswrapper[4726]: I1001 16:01:26.807522 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:26 crc kubenswrapper[4726]: E1001 16:01:26.807707 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:26 crc kubenswrapper[4726]: I1001 16:01:26.807768 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:26 crc kubenswrapper[4726]: E1001 16:01:26.807911 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:26 crc kubenswrapper[4726]: E1001 16:01:26.808088 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:27 crc kubenswrapper[4726]: I1001 16:01:27.810230 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:27 crc kubenswrapper[4726]: E1001 16:01:27.810548 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:27 crc kubenswrapper[4726]: I1001 16:01:27.811895 4726 scope.go:117] "RemoveContainer" containerID="04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be" Oct 01 16:01:27 crc kubenswrapper[4726]: E1001 16:01:27.930802 4726 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 16:01:28 crc kubenswrapper[4726]: I1001 16:01:28.521429 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/3.log" Oct 01 16:01:28 crc kubenswrapper[4726]: I1001 16:01:28.525045 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerStarted","Data":"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8"} Oct 01 16:01:28 crc kubenswrapper[4726]: I1001 16:01:28.525420 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 16:01:28 crc kubenswrapper[4726]: I1001 16:01:28.559883 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podStartSLOduration=118.559867685 podStartE2EDuration="1m58.559867685s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:28.558529929 +0000 UTC m=+141.460082536" watchObservedRunningTime="2025-10-01 16:01:28.559867685 +0000 UTC m=+141.461420262" Oct 01 16:01:28 crc kubenswrapper[4726]: I1001 16:01:28.807495 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:28 crc kubenswrapper[4726]: I1001 16:01:28.807506 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:28 crc kubenswrapper[4726]: E1001 16:01:28.807672 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:28 crc kubenswrapper[4726]: I1001 16:01:28.807522 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:28 crc kubenswrapper[4726]: E1001 16:01:28.807771 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:28 crc kubenswrapper[4726]: E1001 16:01:28.807813 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:28 crc kubenswrapper[4726]: I1001 16:01:28.897777 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4pzd8"] Oct 01 16:01:29 crc kubenswrapper[4726]: I1001 16:01:29.528786 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:29 crc kubenswrapper[4726]: E1001 16:01:29.528940 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:29 crc kubenswrapper[4726]: I1001 16:01:29.807507 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:29 crc kubenswrapper[4726]: E1001 16:01:29.807936 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:30 crc kubenswrapper[4726]: I1001 16:01:30.807733 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:30 crc kubenswrapper[4726]: I1001 16:01:30.807831 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:30 crc kubenswrapper[4726]: I1001 16:01:30.807874 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:30 crc kubenswrapper[4726]: E1001 16:01:30.808023 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:30 crc kubenswrapper[4726]: E1001 16:01:30.808167 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:30 crc kubenswrapper[4726]: E1001 16:01:30.808253 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:31 crc kubenswrapper[4726]: I1001 16:01:31.807706 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:31 crc kubenswrapper[4726]: E1001 16:01:31.807879 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 16:01:32 crc kubenswrapper[4726]: I1001 16:01:32.807232 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:32 crc kubenswrapper[4726]: I1001 16:01:32.807371 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:32 crc kubenswrapper[4726]: I1001 16:01:32.807253 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:32 crc kubenswrapper[4726]: E1001 16:01:32.807572 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 16:01:32 crc kubenswrapper[4726]: E1001 16:01:32.807681 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4pzd8" podUID="2a22e7a8-726a-4538-a27d-44c58f89f0b1" Oct 01 16:01:32 crc kubenswrapper[4726]: E1001 16:01:32.807794 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.807484 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.810072 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.810359 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.901176 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.947299 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl"] Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.947780 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.948341 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mpfpx"] Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.949349 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.949368 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-vxgvr"] Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.952301 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.953849 4726 reflector.go:561] object-"openshift-machine-api"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Oct 01 16:01:33 crc kubenswrapper[4726]: E1001 16:01:33.953921 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.959721 4726 reflector.go:561] object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff": failed to list *v1.Secret: secrets "openshift-apiserver-sa-dockercfg-djjff" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 01 16:01:33 crc kubenswrapper[4726]: E1001 16:01:33.959782 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"openshift-apiserver-sa-dockercfg-djjff\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-apiserver-sa-dockercfg-djjff\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.959853 4726 reflector.go:561] object-"openshift-apiserver"/"image-import-ca": failed to list *v1.ConfigMap: configmaps "image-import-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 01 16:01:33 crc kubenswrapper[4726]: E1001 16:01:33.959873 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"image-import-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"image-import-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.960012 4726 reflector.go:561] object-"openshift-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 01 16:01:33 crc kubenswrapper[4726]: E1001 16:01:33.960039 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.960126 4726 reflector.go:561] object-"openshift-apiserver"/"encryption-config-1": failed to list *v1.Secret: secrets "encryption-config-1" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 01 16:01:33 crc kubenswrapper[4726]: E1001 16:01:33.960147 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"encryption-config-1\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"encryption-config-1\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.960230 4726 reflector.go:561] object-"openshift-apiserver"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 01 16:01:33 crc kubenswrapper[4726]: E1001 16:01:33.960258 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.960345 4726 reflector.go:561] object-"openshift-apiserver"/"etcd-serving-ca": failed to list *v1.ConfigMap: configmaps "etcd-serving-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.960345 4726 reflector.go:561] object-"openshift-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 01 16:01:33 crc kubenswrapper[4726]: E1001 16:01:33.960372 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-serving-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"etcd-serving-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:33 crc kubenswrapper[4726]: E1001 16:01:33.960411 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.960455 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.960735 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.960737 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.960819 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.960899 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.960990 4726 reflector.go:561] object-"openshift-apiserver"/"audit-1": failed to list *v1.ConfigMap: configmaps "audit-1" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 01 16:01:33 crc kubenswrapper[4726]: E1001 16:01:33.961020 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"audit-1\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit-1\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.961108 4726 reflector.go:561] object-"openshift-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 01 16:01:33 crc kubenswrapper[4726]: E1001 16:01:33.961134 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.961335 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.962258 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.962465 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.964165 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.966648 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pxlbv"] Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.967734 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.968484 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8"] Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.969032 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.970212 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cg79v"] Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.971012 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.975670 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.975968 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.976173 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.976305 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.976558 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.977126 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.977309 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.977420 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.977586 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.977617 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.977885 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.978738 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.980145 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw"] Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.980202 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.980610 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.980990 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr"] Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.981871 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.983044 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw" Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.981002 4726 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-images": failed to list *v1.ConfigMap: configmaps "machine-api-operator-images" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Oct 01 16:01:33 crc kubenswrapper[4726]: E1001 16:01:33.983487 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-images\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"machine-api-operator-images\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.982580 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.982622 4726 reflector.go:561] object-"openshift-apiserver"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 01 16:01:33 crc kubenswrapper[4726]: E1001 16:01:33.983644 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.982676 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.982728 4726 reflector.go:561] object-"openshift-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 01 16:01:33 crc kubenswrapper[4726]: E1001 16:01:33.983713 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.982765 4726 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7": failed to list *v1.Secret: secrets "machine-api-operator-dockercfg-mfbb7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Oct 01 16:01:33 crc kubenswrapper[4726]: E1001 16:01:33.983732 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-mfbb7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-dockercfg-mfbb7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.987208 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7jshc"] Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.987616 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zhfrh"] Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.987952 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.988095 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.988222 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79"] Oct 01 16:01:33 crc kubenswrapper[4726]: I1001 16:01:33.988912 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79" Oct 01 16:01:33 crc kubenswrapper[4726]: W1001 16:01:33.992125 4726 reflector.go:561] object-"openshift-machine-api"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:33.999902 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 01 16:01:34 crc kubenswrapper[4726]: W1001 16:01:33.992158 4726 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-tls": failed to list *v1.Secret: secrets "machine-api-operator-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Oct 01 16:01:34 crc kubenswrapper[4726]: W1001 16:01:33.995196 4726 reflector.go:561] object-"openshift-machine-api"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:33.998410 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:33.998453 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:33.998486 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:33.998532 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:33.998590 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:33.998714 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:33.998883 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:33.998991 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:33.999022 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.000385 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 01 16:01:34 crc kubenswrapper[4726]: E1001 16:01:33.999911 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:34 crc kubenswrapper[4726]: E1001 16:01:34.000861 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:34 crc kubenswrapper[4726]: E1001 16:01:34.000993 4726 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:33.994131 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-w4c9z"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.001578 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.001683 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.001839 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.001888 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pq5pm"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.001971 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.002181 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.002235 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.002437 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.002767 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.003160 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-5tjmt"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.003605 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.003887 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z4ml2"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.003938 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.004234 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z4ml2" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.004468 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-5tjmt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.004491 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.003890 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.004748 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.004987 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.005861 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.006591 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.012167 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.012343 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.012471 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.030293 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.032313 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.032436 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.033888 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-etcd-client\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.033914 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-audit-policies\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.033943 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhwlw\" (UniqueName: \"kubernetes.io/projected/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-kube-api-access-jhwlw\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.033996 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-serving-cert\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034026 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-node-pullsecrets\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034042 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-audit\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034125 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034148 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5399581d-2111-47d1-bfe7-007b7c46d706-images\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034172 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-encryption-config\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034231 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5399581d-2111-47d1-bfe7-007b7c46d706-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034250 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-encryption-config\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034271 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkll2\" (UniqueName: \"kubernetes.io/projected/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-kube-api-access-hkll2\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034335 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5399581d-2111-47d1-bfe7-007b7c46d706-config\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034351 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034381 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-audit-dir\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034396 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-etcd-client\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034445 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-audit-dir\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034463 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52fd2\" (UniqueName: \"kubernetes.io/projected/5399581d-2111-47d1-bfe7-007b7c46d706-kube-api-access-52fd2\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034483 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-config\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034519 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-image-import-ca\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034535 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-etcd-serving-ca\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034564 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.034584 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-serving-cert\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.035076 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.035649 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.035703 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.035980 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.036126 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.036336 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.036136 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.036342 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.036672 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.036721 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.037015 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.054616 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.055985 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.056006 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.056149 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j9fr5"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.056254 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.056494 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.056651 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.056696 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.056733 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.056809 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.056906 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.057009 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.057123 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.057156 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.057020 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.057230 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.057337 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.057386 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.057425 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.057440 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.057543 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.057976 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.058431 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.058806 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.059573 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.059876 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.060277 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-np6df"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.060569 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.060885 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.061370 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.061612 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-np6df" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.062183 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-9x846"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.062659 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.064813 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.064900 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.064938 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6w44d"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.065618 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6w44d" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.067471 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.069741 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.070287 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.072657 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.072823 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mpfpx"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.074480 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.088202 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.097367 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.100099 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.100836 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.100923 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.101462 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.101713 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.102984 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.115292 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4r5zz"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.115849 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-4r5zz" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.119039 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.119468 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qvwb"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.119794 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.119829 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.120248 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.120671 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.121307 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.122519 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.122904 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.130210 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.131019 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.132559 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.133631 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.134100 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.134747 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-b842d"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135226 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135454 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135493 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16ec2ea5-2c69-457f-b36f-06011c98666e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-b62jx\" (UID: \"16ec2ea5-2c69-457f-b36f-06011c98666e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135521 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135538 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135558 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f189676-cdd8-4138-b73f-863e1c075bc2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rgfgr\" (UID: \"7f189676-cdd8-4138-b73f-863e1c075bc2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135591 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f189676-cdd8-4138-b73f-863e1c075bc2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rgfgr\" (UID: \"7f189676-cdd8-4138-b73f-863e1c075bc2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135608 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-serving-cert\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135624 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/51675697-3e42-4381-a278-44377e79479e-metrics-tls\") pod \"dns-operator-744455d44c-np6df\" (UID: \"51675697-3e42-4381-a278-44377e79479e\") " pod="openshift-dns-operator/dns-operator-744455d44c-np6df" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135638 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135661 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135676 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135691 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6jp9\" (UniqueName: \"kubernetes.io/projected/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-kube-api-access-h6jp9\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135707 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzk4t\" (UniqueName: \"kubernetes.io/projected/7f189676-cdd8-4138-b73f-863e1c075bc2-kube-api-access-vzk4t\") pod \"openshift-apiserver-operator-796bbdcf4f-rgfgr\" (UID: \"7f189676-cdd8-4138-b73f-863e1c075bc2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135723 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-audit-policies\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135736 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-etcd-client\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135752 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135767 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25ce2f03-84c1-496b-9dbc-c8efba41e3d3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2sgp2\" (UID: \"25ce2f03-84c1-496b-9dbc-c8efba41e3d3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135784 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhwlw\" (UniqueName: \"kubernetes.io/projected/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-kube-api-access-jhwlw\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135798 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/654d70bb-a9aa-48d1-99dd-95a50b221440-trusted-ca\") pod \"console-operator-58897d9998-7jshc\" (UID: \"654d70bb-a9aa-48d1-99dd-95a50b221440\") " pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135814 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/25ce2f03-84c1-496b-9dbc-c8efba41e3d3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2sgp2\" (UID: \"25ce2f03-84c1-496b-9dbc-c8efba41e3d3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135829 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/16ec2ea5-2c69-457f-b36f-06011c98666e-metrics-tls\") pod \"ingress-operator-5b745b69d9-b62jx\" (UID: \"16ec2ea5-2c69-457f-b36f-06011c98666e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135844 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-serving-cert\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135859 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27mwc\" (UniqueName: \"kubernetes.io/projected/73789e0a-a83d-4470-9f4c-137b3c317b85-kube-api-access-27mwc\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135882 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135904 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-node-pullsecrets\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135919 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-audit\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135935 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/25ce2f03-84c1-496b-9dbc-c8efba41e3d3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2sgp2\" (UID: \"25ce2f03-84c1-496b-9dbc-c8efba41e3d3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135953 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-config\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135969 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.135987 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-encryption-config\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136001 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5399581d-2111-47d1-bfe7-007b7c46d706-images\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136016 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136031 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q8b7\" (UniqueName: \"kubernetes.io/projected/25ce2f03-84c1-496b-9dbc-c8efba41e3d3-kube-api-access-5q8b7\") pod \"cluster-image-registry-operator-dc59b4c8b-2sgp2\" (UID: \"25ce2f03-84c1-496b-9dbc-c8efba41e3d3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136060 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-audit-dir\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136075 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136089 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f62rz\" (UniqueName: \"kubernetes.io/projected/654d70bb-a9aa-48d1-99dd-95a50b221440-kube-api-access-f62rz\") pod \"console-operator-58897d9998-7jshc\" (UID: \"654d70bb-a9aa-48d1-99dd-95a50b221440\") " pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136106 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5399581d-2111-47d1-bfe7-007b7c46d706-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136120 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn9n7\" (UniqueName: \"kubernetes.io/projected/16ec2ea5-2c69-457f-b36f-06011c98666e-kube-api-access-wn9n7\") pod \"ingress-operator-5b745b69d9-b62jx\" (UID: \"16ec2ea5-2c69-457f-b36f-06011c98666e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136136 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73789e0a-a83d-4470-9f4c-137b3c317b85-serving-cert\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136150 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/654d70bb-a9aa-48d1-99dd-95a50b221440-serving-cert\") pod \"console-operator-58897d9998-7jshc\" (UID: \"654d70bb-a9aa-48d1-99dd-95a50b221440\") " pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136165 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-encryption-config\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136182 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f008910-43f3-4260-86e6-4cd5a92e28c0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-67558\" (UID: \"2f008910-43f3-4260-86e6-4cd5a92e28c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136196 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkll2\" (UniqueName: \"kubernetes.io/projected/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-kube-api-access-hkll2\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136220 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136236 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136251 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5399581d-2111-47d1-bfe7-007b7c46d706-config\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136264 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-client-ca\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136279 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f008910-43f3-4260-86e6-4cd5a92e28c0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-67558\" (UID: \"2f008910-43f3-4260-86e6-4cd5a92e28c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136295 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f008910-43f3-4260-86e6-4cd5a92e28c0-config\") pod \"kube-controller-manager-operator-78b949d7b-67558\" (UID: \"2f008910-43f3-4260-86e6-4cd5a92e28c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136320 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-audit-dir\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136335 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-etcd-client\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136350 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-audit-policies\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136367 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj5ds\" (UniqueName: \"kubernetes.io/projected/51675697-3e42-4381-a278-44377e79479e-kube-api-access-nj5ds\") pod \"dns-operator-744455d44c-np6df\" (UID: \"51675697-3e42-4381-a278-44377e79479e\") " pod="openshift-dns-operator/dns-operator-744455d44c-np6df" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136384 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-audit-dir\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136398 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52fd2\" (UniqueName: \"kubernetes.io/projected/5399581d-2111-47d1-bfe7-007b7c46d706-kube-api-access-52fd2\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136413 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/654d70bb-a9aa-48d1-99dd-95a50b221440-config\") pod \"console-operator-58897d9998-7jshc\" (UID: \"654d70bb-a9aa-48d1-99dd-95a50b221440\") " pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136428 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16ec2ea5-2c69-457f-b36f-06011c98666e-trusted-ca\") pod \"ingress-operator-5b745b69d9-b62jx\" (UID: \"16ec2ea5-2c69-457f-b36f-06011c98666e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136444 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-config\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136458 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-image-import-ca\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136474 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136490 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.136507 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-etcd-serving-ca\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.137021 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.137270 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7rq8t"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.137876 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-node-pullsecrets\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.137992 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pxlbv"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.138091 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rq8t" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.138640 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-audit-dir\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.138772 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-audit-dir\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.139070 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-audit-policies\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.139299 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.139515 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-5jsl4"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.140035 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-5jsl4" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.142636 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cg79v"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.142687 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.143295 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.144231 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.145178 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j9fr5"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.146339 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-w4c9z"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.146486 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-encryption-config\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.146649 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-serving-cert\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.147673 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-vxgvr"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.150084 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.150280 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6w44d"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.150313 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7jshc"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.150634 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-etcd-client\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.151059 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.153436 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.153966 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.157894 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.160643 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.162760 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pq5pm"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.166004 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.167699 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.169475 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.170098 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.170545 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-5tjmt"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.175183 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-l6d7x"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.176184 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-l6d7x" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.178039 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-j96mp"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.179309 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.179859 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-np6df"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.181403 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zhfrh"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.184922 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qvwb"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.186566 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.188680 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.189909 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.190554 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.191882 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.192556 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z4ml2"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.193668 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-j96mp"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.195085 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4r5zz"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.201667 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.207940 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7rq8t"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.209331 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.210266 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-5jsl4"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.211417 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.212969 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-l6d7x"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.214740 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-b842d"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.216363 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.217788 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-6p58q"] Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.218410 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6p58q" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.229577 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.236989 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16ec2ea5-2c69-457f-b36f-06011c98666e-trusted-ca\") pod \"ingress-operator-5b745b69d9-b62jx\" (UID: \"16ec2ea5-2c69-457f-b36f-06011c98666e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.237235 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/654d70bb-a9aa-48d1-99dd-95a50b221440-config\") pod \"console-operator-58897d9998-7jshc\" (UID: \"654d70bb-a9aa-48d1-99dd-95a50b221440\") " pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.237314 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.237412 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.237532 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16ec2ea5-2c69-457f-b36f-06011c98666e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-b62jx\" (UID: \"16ec2ea5-2c69-457f-b36f-06011c98666e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.237654 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.237761 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.237867 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f189676-cdd8-4138-b73f-863e1c075bc2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rgfgr\" (UID: \"7f189676-cdd8-4138-b73f-863e1c075bc2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.237962 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f189676-cdd8-4138-b73f-863e1c075bc2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rgfgr\" (UID: \"7f189676-cdd8-4138-b73f-863e1c075bc2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.238034 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/51675697-3e42-4381-a278-44377e79479e-metrics-tls\") pod \"dns-operator-744455d44c-np6df\" (UID: \"51675697-3e42-4381-a278-44377e79479e\") " pod="openshift-dns-operator/dns-operator-744455d44c-np6df" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.238133 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.238209 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.238284 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.238356 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6jp9\" (UniqueName: \"kubernetes.io/projected/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-kube-api-access-h6jp9\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.238438 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzk4t\" (UniqueName: \"kubernetes.io/projected/7f189676-cdd8-4138-b73f-863e1c075bc2-kube-api-access-vzk4t\") pod \"openshift-apiserver-operator-796bbdcf4f-rgfgr\" (UID: \"7f189676-cdd8-4138-b73f-863e1c075bc2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.238524 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.238628 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25ce2f03-84c1-496b-9dbc-c8efba41e3d3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2sgp2\" (UID: \"25ce2f03-84c1-496b-9dbc-c8efba41e3d3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.238717 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/654d70bb-a9aa-48d1-99dd-95a50b221440-trusted-ca\") pod \"console-operator-58897d9998-7jshc\" (UID: \"654d70bb-a9aa-48d1-99dd-95a50b221440\") " pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.238791 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/25ce2f03-84c1-496b-9dbc-c8efba41e3d3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2sgp2\" (UID: \"25ce2f03-84c1-496b-9dbc-c8efba41e3d3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.238850 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f189676-cdd8-4138-b73f-863e1c075bc2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rgfgr\" (UID: \"7f189676-cdd8-4138-b73f-863e1c075bc2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.238864 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/16ec2ea5-2c69-457f-b36f-06011c98666e-metrics-tls\") pod \"ingress-operator-5b745b69d9-b62jx\" (UID: \"16ec2ea5-2c69-457f-b36f-06011c98666e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.238966 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27mwc\" (UniqueName: \"kubernetes.io/projected/73789e0a-a83d-4470-9f4c-137b3c317b85-kube-api-access-27mwc\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.239002 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.239076 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/25ce2f03-84c1-496b-9dbc-c8efba41e3d3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2sgp2\" (UID: \"25ce2f03-84c1-496b-9dbc-c8efba41e3d3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.239118 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.239127 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.239231 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-config\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.238905 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.239318 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.239503 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q8b7\" (UniqueName: \"kubernetes.io/projected/25ce2f03-84c1-496b-9dbc-c8efba41e3d3-kube-api-access-5q8b7\") pod \"cluster-image-registry-operator-dc59b4c8b-2sgp2\" (UID: \"25ce2f03-84c1-496b-9dbc-c8efba41e3d3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.239546 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.239576 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f62rz\" (UniqueName: \"kubernetes.io/projected/654d70bb-a9aa-48d1-99dd-95a50b221440-kube-api-access-f62rz\") pod \"console-operator-58897d9998-7jshc\" (UID: \"654d70bb-a9aa-48d1-99dd-95a50b221440\") " pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.239692 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-audit-dir\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.239931 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn9n7\" (UniqueName: \"kubernetes.io/projected/16ec2ea5-2c69-457f-b36f-06011c98666e-kube-api-access-wn9n7\") pod \"ingress-operator-5b745b69d9-b62jx\" (UID: \"16ec2ea5-2c69-457f-b36f-06011c98666e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.239958 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73789e0a-a83d-4470-9f4c-137b3c317b85-serving-cert\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.239979 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/654d70bb-a9aa-48d1-99dd-95a50b221440-serving-cert\") pod \"console-operator-58897d9998-7jshc\" (UID: \"654d70bb-a9aa-48d1-99dd-95a50b221440\") " pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.239977 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/654d70bb-a9aa-48d1-99dd-95a50b221440-trusted-ca\") pod \"console-operator-58897d9998-7jshc\" (UID: \"654d70bb-a9aa-48d1-99dd-95a50b221440\") " pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.240011 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f008910-43f3-4260-86e6-4cd5a92e28c0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-67558\" (UID: \"2f008910-43f3-4260-86e6-4cd5a92e28c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.240012 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25ce2f03-84c1-496b-9dbc-c8efba41e3d3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2sgp2\" (UID: \"25ce2f03-84c1-496b-9dbc-c8efba41e3d3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.240080 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.240127 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-client-ca\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.240155 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f008910-43f3-4260-86e6-4cd5a92e28c0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-67558\" (UID: \"2f008910-43f3-4260-86e6-4cd5a92e28c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.240190 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f008910-43f3-4260-86e6-4cd5a92e28c0-config\") pod \"kube-controller-manager-operator-78b949d7b-67558\" (UID: \"2f008910-43f3-4260-86e6-4cd5a92e28c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.240226 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-audit-policies\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.240253 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj5ds\" (UniqueName: \"kubernetes.io/projected/51675697-3e42-4381-a278-44377e79479e-kube-api-access-nj5ds\") pod \"dns-operator-744455d44c-np6df\" (UID: \"51675697-3e42-4381-a278-44377e79479e\") " pod="openshift-dns-operator/dns-operator-744455d44c-np6df" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.240387 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16ec2ea5-2c69-457f-b36f-06011c98666e-trusted-ca\") pod \"ingress-operator-5b745b69d9-b62jx\" (UID: \"16ec2ea5-2c69-457f-b36f-06011c98666e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.240570 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-audit-dir\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.241158 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.241198 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-config\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.242281 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-client-ca\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.242613 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.242927 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/25ce2f03-84c1-496b-9dbc-c8efba41e3d3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2sgp2\" (UID: \"25ce2f03-84c1-496b-9dbc-c8efba41e3d3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.243135 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/654d70bb-a9aa-48d1-99dd-95a50b221440-config\") pod \"console-operator-58897d9998-7jshc\" (UID: \"654d70bb-a9aa-48d1-99dd-95a50b221440\") " pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.243177 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-audit-policies\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.243250 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.244079 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/654d70bb-a9aa-48d1-99dd-95a50b221440-serving-cert\") pod \"console-operator-58897d9998-7jshc\" (UID: \"654d70bb-a9aa-48d1-99dd-95a50b221440\") " pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.244677 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.244914 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f189676-cdd8-4138-b73f-863e1c075bc2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rgfgr\" (UID: \"7f189676-cdd8-4138-b73f-863e1c075bc2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.245027 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.245171 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.245424 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.245587 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.245712 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73789e0a-a83d-4470-9f4c-137b3c317b85-serving-cert\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.246215 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.250104 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.270079 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.282712 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/16ec2ea5-2c69-457f-b36f-06011c98666e-metrics-tls\") pod \"ingress-operator-5b745b69d9-b62jx\" (UID: \"16ec2ea5-2c69-457f-b36f-06011c98666e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.290575 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.309773 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.315342 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f008910-43f3-4260-86e6-4cd5a92e28c0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-67558\" (UID: \"2f008910-43f3-4260-86e6-4cd5a92e28c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.329621 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.332686 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f008910-43f3-4260-86e6-4cd5a92e28c0-config\") pod \"kube-controller-manager-operator-78b949d7b-67558\" (UID: \"2f008910-43f3-4260-86e6-4cd5a92e28c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.349890 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.369844 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.390352 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.409778 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.430124 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.450513 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.470027 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.482107 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/51675697-3e42-4381-a278-44377e79479e-metrics-tls\") pod \"dns-operator-744455d44c-np6df\" (UID: \"51675697-3e42-4381-a278-44377e79479e\") " pod="openshift-dns-operator/dns-operator-744455d44c-np6df" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.489853 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.509840 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.530584 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.549825 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.570081 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.589816 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.609643 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.629977 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.650454 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.669836 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.690763 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.711152 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.771245 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.790689 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.807389 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.807414 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.807521 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.811378 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.830134 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.850972 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.879044 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.890018 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.910466 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.930098 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.950099 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.970437 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 01 16:01:34 crc kubenswrapper[4726]: I1001 16:01:34.990344 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.010890 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.030423 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.050454 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.069850 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.091170 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.110181 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.128495 4726 request.go:700] Waited for 1.012300036s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca/secrets?fieldSelector=metadata.name%3Dsigning-key&limit=500&resourceVersion=0 Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.130763 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.136706 4726 configmap.go:193] Couldn't get configMap openshift-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.136820 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-etcd-serving-ca podName:e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f nodeName:}" failed. No retries permitted until 2025-10-01 16:01:35.636789848 +0000 UTC m=+148.538342455 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-etcd-serving-ca") pod "apiserver-76f77b778f-vxgvr" (UID: "e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f") : failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138027 4726 secret.go:188] Couldn't get secret openshift-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138122 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-serving-cert podName:e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f nodeName:}" failed. No retries permitted until 2025-10-01 16:01:35.638102863 +0000 UTC m=+148.539655480 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-serving-cert") pod "apiserver-76f77b778f-vxgvr" (UID: "e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f") : failed to sync secret cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138151 4726 configmap.go:193] Couldn't get configMap openshift-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138248 4726 configmap.go:193] Couldn't get configMap openshift-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138385 4726 configmap.go:193] Couldn't get configMap openshift-machine-api/machine-api-operator-images: failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138393 4726 secret.go:188] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138334 4726 secret.go:188] Couldn't get secret openshift-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138288 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-trusted-ca-bundle podName:e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f nodeName:}" failed. No retries permitted until 2025-10-01 16:01:35.638248327 +0000 UTC m=+148.539801034 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-trusted-ca-bundle") pod "apiserver-76f77b778f-vxgvr" (UID: "e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f") : failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138693 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-audit podName:e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f nodeName:}" failed. No retries permitted until 2025-10-01 16:01:35.638649397 +0000 UTC m=+148.540202034 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit" (UniqueName: "kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-audit") pod "apiserver-76f77b778f-vxgvr" (UID: "e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f") : failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138710 4726 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138746 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5399581d-2111-47d1-bfe7-007b7c46d706-images podName:5399581d-2111-47d1-bfe7-007b7c46d706 nodeName:}" failed. No retries permitted until 2025-10-01 16:01:35.638727629 +0000 UTC m=+148.540280346 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/5399581d-2111-47d1-bfe7-007b7c46d706-images") pod "machine-api-operator-5694c8668f-mpfpx" (UID: "5399581d-2111-47d1-bfe7-007b7c46d706") : failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138789 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5399581d-2111-47d1-bfe7-007b7c46d706-config podName:5399581d-2111-47d1-bfe7-007b7c46d706 nodeName:}" failed. No retries permitted until 2025-10-01 16:01:35.63876387 +0000 UTC m=+148.540316587 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/5399581d-2111-47d1-bfe7-007b7c46d706-config") pod "machine-api-operator-5694c8668f-mpfpx" (UID: "5399581d-2111-47d1-bfe7-007b7c46d706") : failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138811 4726 configmap.go:193] Couldn't get configMap openshift-apiserver/config: failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138827 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5399581d-2111-47d1-bfe7-007b7c46d706-machine-api-operator-tls podName:5399581d-2111-47d1-bfe7-007b7c46d706 nodeName:}" failed. No retries permitted until 2025-10-01 16:01:35.638809532 +0000 UTC m=+148.540362299 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/5399581d-2111-47d1-bfe7-007b7c46d706-machine-api-operator-tls") pod "machine-api-operator-5694c8668f-mpfpx" (UID: "5399581d-2111-47d1-bfe7-007b7c46d706") : failed to sync secret cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138841 4726 configmap.go:193] Couldn't get configMap openshift-apiserver/image-import-ca: failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138857 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-encryption-config podName:e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f nodeName:}" failed. No retries permitted until 2025-10-01 16:01:35.638841082 +0000 UTC m=+148.540393829 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-encryption-config") pod "apiserver-76f77b778f-vxgvr" (UID: "e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f") : failed to sync secret cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138885 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-config podName:e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f nodeName:}" failed. No retries permitted until 2025-10-01 16:01:35.638871563 +0000 UTC m=+148.540424290 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-config") pod "apiserver-76f77b778f-vxgvr" (UID: "e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f") : failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138893 4726 secret.go:188] Couldn't get secret openshift-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.138916 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-image-import-ca podName:e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f nodeName:}" failed. No retries permitted until 2025-10-01 16:01:35.638901464 +0000 UTC m=+148.540454181 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "image-import-ca" (UniqueName: "kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-image-import-ca") pod "apiserver-76f77b778f-vxgvr" (UID: "e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f") : failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: E1001 16:01:35.139116 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-etcd-client podName:e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f nodeName:}" failed. No retries permitted until 2025-10-01 16:01:35.638995517 +0000 UTC m=+148.540548184 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-etcd-client") pod "apiserver-76f77b778f-vxgvr" (UID: "e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f") : failed to sync secret cache: timed out waiting for the condition Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.151951 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.170649 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.191670 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.210440 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.230702 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.249978 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.276409 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.289104 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.310003 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.331017 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.350402 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.371350 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.390165 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.410913 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.430586 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.449983 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.470156 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.490359 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.510424 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.530458 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.550032 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.570748 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.590572 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.609750 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.629694 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.649992 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.657704 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.657769 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5399581d-2111-47d1-bfe7-007b7c46d706-images\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.657838 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5399581d-2111-47d1-bfe7-007b7c46d706-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.657884 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-encryption-config\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.657960 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5399581d-2111-47d1-bfe7-007b7c46d706-config\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.658021 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-etcd-client\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.658164 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-config\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.658249 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-image-import-ca\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.658298 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-etcd-serving-ca\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.658443 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-serving-cert\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.658512 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-audit\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.669458 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.710207 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.730538 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.750322 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.795575 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkll2\" (UniqueName: \"kubernetes.io/projected/262d42bb-7bb1-49f1-8cf0-ad225e6dd673-kube-api-access-hkll2\") pod \"apiserver-7bbb656c7d-8xtnl\" (UID: \"262d42bb-7bb1-49f1-8cf0-ad225e6dd673\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.809710 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.831316 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.850092 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.869602 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.890562 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.910442 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.930351 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.950150 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.970928 4726 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 01 16:01:35 crc kubenswrapper[4726]: I1001 16:01:35.991388 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.010646 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.030790 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.050254 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.086528 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.104175 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16ec2ea5-2c69-457f-b36f-06011c98666e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-b62jx\" (UID: \"16ec2ea5-2c69-457f-b36f-06011c98666e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.115564 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6jp9\" (UniqueName: \"kubernetes.io/projected/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-kube-api-access-h6jp9\") pod \"oauth-openshift-558db77b4-pq5pm\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.128716 4726 request.go:700] Waited for 1.889409635s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/serviceaccounts/cluster-image-registry-operator/token Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.145652 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/25ce2f03-84c1-496b-9dbc-c8efba41e3d3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2sgp2\" (UID: \"25ce2f03-84c1-496b-9dbc-c8efba41e3d3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.145707 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzk4t\" (UniqueName: \"kubernetes.io/projected/7f189676-cdd8-4138-b73f-863e1c075bc2-kube-api-access-vzk4t\") pod \"openshift-apiserver-operator-796bbdcf4f-rgfgr\" (UID: \"7f189676-cdd8-4138-b73f-863e1c075bc2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.166946 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27mwc\" (UniqueName: \"kubernetes.io/projected/73789e0a-a83d-4470-9f4c-137b3c317b85-kube-api-access-27mwc\") pod \"controller-manager-879f6c89f-zhfrh\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.184314 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj5ds\" (UniqueName: \"kubernetes.io/projected/51675697-3e42-4381-a278-44377e79479e-kube-api-access-nj5ds\") pod \"dns-operator-744455d44c-np6df\" (UID: \"51675697-3e42-4381-a278-44377e79479e\") " pod="openshift-dns-operator/dns-operator-744455d44c-np6df" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.205015 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q8b7\" (UniqueName: \"kubernetes.io/projected/25ce2f03-84c1-496b-9dbc-c8efba41e3d3-kube-api-access-5q8b7\") pod \"cluster-image-registry-operator-dc59b4c8b-2sgp2\" (UID: \"25ce2f03-84c1-496b-9dbc-c8efba41e3d3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.224487 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.237084 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f008910-43f3-4260-86e6-4cd5a92e28c0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-67558\" (UID: \"2f008910-43f3-4260-86e6-4cd5a92e28c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.247182 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.255746 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f62rz\" (UniqueName: \"kubernetes.io/projected/654d70bb-a9aa-48d1-99dd-95a50b221440-kube-api-access-f62rz\") pod \"console-operator-58897d9998-7jshc\" (UID: \"654d70bb-a9aa-48d1-99dd-95a50b221440\") " pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.267458 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn9n7\" (UniqueName: \"kubernetes.io/projected/16ec2ea5-2c69-457f-b36f-06011c98666e-kube-api-access-wn9n7\") pod \"ingress-operator-5b745b69d9-b62jx\" (UID: \"16ec2ea5-2c69-457f-b36f-06011c98666e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.274579 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.310573 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.325482 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl"] Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.325654 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.332442 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.337460 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.350426 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.369765 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a8d300c-d03d-4927-b9ba-c17f692a9db4-serving-cert\") pod \"route-controller-manager-6576b87f9c-2gcb8\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.369820 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a8d300c-d03d-4927-b9ba-c17f692a9db4-config\") pod \"route-controller-manager-6576b87f9c-2gcb8\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.369858 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcp85\" (UniqueName: \"kubernetes.io/projected/8282f47b-cd59-4cba-a1aa-7528d57842ec-kube-api-access-qcp85\") pod \"control-plane-machine-set-operator-78cbb6b69f-z4ml2\" (UID: \"8282f47b-cd59-4cba-a1aa-7528d57842ec\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z4ml2" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.369881 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10a86730-3fae-4053-b4e0-48effedbf3de-proxy-tls\") pod \"machine-config-operator-74547568cd-xh5hv\" (UID: \"10a86730-3fae-4053-b4e0-48effedbf3de\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.369901 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a44f6721-0fe0-45cc-8290-fe52ffb3f8f2-serving-cert\") pod \"openshift-config-operator-7777fb866f-cg79v\" (UID: \"a44f6721-0fe0-45cc-8290-fe52ffb3f8f2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.369923 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6k29\" (UniqueName: \"kubernetes.io/projected/e500be09-9f04-49a8-8ad2-e69012bd36f1-kube-api-access-d6k29\") pod \"multus-admission-controller-857f4d67dd-6w44d\" (UID: \"e500be09-9f04-49a8-8ad2-e69012bd36f1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6w44d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.369946 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dns5\" (UniqueName: \"kubernetes.io/projected/b8cf6941-f833-4465-a3e9-9efea4b9b5b0-kube-api-access-7dns5\") pod \"downloads-7954f5f757-5tjmt\" (UID: \"b8cf6941-f833-4465-a3e9-9efea4b9b5b0\") " pod="openshift-console/downloads-7954f5f757-5tjmt" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.369969 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6zb9\" (UniqueName: \"kubernetes.io/projected/a44f6721-0fe0-45cc-8290-fe52ffb3f8f2-kube-api-access-b6zb9\") pod \"openshift-config-operator-7777fb866f-cg79v\" (UID: \"a44f6721-0fe0-45cc-8290-fe52ffb3f8f2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.369996 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f188f7e-b855-4902-b172-db25f2a53746-service-ca-bundle\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370023 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/496e7e53-b6f6-48b1-9657-a091d21c301f-trusted-ca\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370045 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhm97\" (UniqueName: \"kubernetes.io/projected/10a86730-3fae-4053-b4e0-48effedbf3de-kube-api-access-rhm97\") pod \"machine-config-operator-74547568cd-xh5hv\" (UID: \"10a86730-3fae-4053-b4e0-48effedbf3de\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370084 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/496e7e53-b6f6-48b1-9657-a091d21c301f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370105 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f188f7e-b855-4902-b172-db25f2a53746-config\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370136 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgpvw\" (UniqueName: \"kubernetes.io/projected/60142257-8641-4fba-83ca-bb1af9f68c0d-kube-api-access-tgpvw\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370157 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a79834ba-7c32-4c45-a927-66cc2ebf3e97-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k6p79\" (UID: \"a79834ba-7c32-4c45-a927-66cc2ebf3e97\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370182 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f371d8f0-5386-4a69-b81c-8c656d71cd54-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-54rxk\" (UID: \"f371d8f0-5386-4a69-b81c-8c656d71cd54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370211 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz2rc\" (UniqueName: \"kubernetes.io/projected/3f188f7e-b855-4902-b172-db25f2a53746-kube-api-access-xz2rc\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370231 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/79d896b5-b34d-4206-92d8-99caeafdffb2-console-serving-cert\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370250 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-service-ca\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370279 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-trusted-ca-bundle\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370302 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt7mt\" (UniqueName: \"kubernetes.io/projected/1a8d300c-d03d-4927-b9ba-c17f692a9db4-kube-api-access-nt7mt\") pod \"route-controller-manager-6576b87f9c-2gcb8\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370346 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f188f7e-b855-4902-b172-db25f2a53746-serving-cert\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370370 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/496e7e53-b6f6-48b1-9657-a091d21c301f-registry-certificates\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370387 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15670655-7e86-4731-bb0b-7ebd96a20e59-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fkxmw\" (UID: \"15670655-7e86-4731-bb0b-7ebd96a20e59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370405 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/052caf82-2622-48b3-8c66-02955a93dcf5-config\") pod \"machine-approver-56656f9798-m4w6w\" (UID: \"052caf82-2622-48b3-8c66-02955a93dcf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370437 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7tsz\" (UniqueName: \"kubernetes.io/projected/15670655-7e86-4731-bb0b-7ebd96a20e59-kube-api-access-g7tsz\") pod \"openshift-controller-manager-operator-756b6f6bc6-fkxmw\" (UID: \"15670655-7e86-4731-bb0b-7ebd96a20e59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370458 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/60142257-8641-4fba-83ca-bb1af9f68c0d-default-certificate\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370478 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-oauth-serving-cert\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370509 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8282f47b-cd59-4cba-a1aa-7528d57842ec-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z4ml2\" (UID: \"8282f47b-cd59-4cba-a1aa-7528d57842ec\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z4ml2" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370533 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82hfx\" (UniqueName: \"kubernetes.io/projected/a79834ba-7c32-4c45-a927-66cc2ebf3e97-kube-api-access-82hfx\") pod \"cluster-samples-operator-665b6dd947-k6p79\" (UID: \"a79834ba-7c32-4c45-a927-66cc2ebf3e97\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370554 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/052caf82-2622-48b3-8c66-02955a93dcf5-auth-proxy-config\") pod \"machine-approver-56656f9798-m4w6w\" (UID: \"052caf82-2622-48b3-8c66-02955a93dcf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370575 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-console-config\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370605 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7smz\" (UniqueName: \"kubernetes.io/projected/052caf82-2622-48b3-8c66-02955a93dcf5-kube-api-access-d7smz\") pod \"machine-approver-56656f9798-m4w6w\" (UID: \"052caf82-2622-48b3-8c66-02955a93dcf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370626 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/10a86730-3fae-4053-b4e0-48effedbf3de-images\") pod \"machine-config-operator-74547568cd-xh5hv\" (UID: \"10a86730-3fae-4053-b4e0-48effedbf3de\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370647 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60142257-8641-4fba-83ca-bb1af9f68c0d-service-ca-bundle\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370669 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60142257-8641-4fba-83ca-bb1af9f68c0d-metrics-certs\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370710 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-bound-sa-token\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370732 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15670655-7e86-4731-bb0b-7ebd96a20e59-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fkxmw\" (UID: \"15670655-7e86-4731-bb0b-7ebd96a20e59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370752 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e500be09-9f04-49a8-8ad2-e69012bd36f1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6w44d\" (UID: \"e500be09-9f04-49a8-8ad2-e69012bd36f1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6w44d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370772 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/60142257-8641-4fba-83ca-bb1af9f68c0d-stats-auth\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370793 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a8d300c-d03d-4927-b9ba-c17f692a9db4-client-ca\") pod \"route-controller-manager-6576b87f9c-2gcb8\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370816 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-registry-tls\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370834 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/052caf82-2622-48b3-8c66-02955a93dcf5-machine-approver-tls\") pod \"machine-approver-56656f9798-m4w6w\" (UID: \"052caf82-2622-48b3-8c66-02955a93dcf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370851 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10a86730-3fae-4053-b4e0-48effedbf3de-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xh5hv\" (UID: \"10a86730-3fae-4053-b4e0-48effedbf3de\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370877 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370911 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwzxw\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-kube-api-access-pwzxw\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370930 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f188f7e-b855-4902-b172-db25f2a53746-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370950 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a44f6721-0fe0-45cc-8290-fe52ffb3f8f2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cg79v\" (UID: \"a44f6721-0fe0-45cc-8290-fe52ffb3f8f2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370970 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgmc4\" (UniqueName: \"kubernetes.io/projected/f371d8f0-5386-4a69-b81c-8c656d71cd54-kube-api-access-wgmc4\") pod \"machine-config-controller-84d6567774-54rxk\" (UID: \"f371d8f0-5386-4a69-b81c-8c656d71cd54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.370988 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/79d896b5-b34d-4206-92d8-99caeafdffb2-console-oauth-config\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.371015 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f371d8f0-5386-4a69-b81c-8c656d71cd54-proxy-tls\") pod \"machine-config-controller-84d6567774-54rxk\" (UID: \"f371d8f0-5386-4a69-b81c-8c656d71cd54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.371031 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x82vr\" (UniqueName: \"kubernetes.io/projected/79d896b5-b34d-4206-92d8-99caeafdffb2-kube-api-access-x82vr\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.371099 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/496e7e53-b6f6-48b1-9657-a091d21c301f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: E1001 16:01:36.371779 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:36.871764688 +0000 UTC m=+149.773317255 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.372284 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.372401 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.383595 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.390350 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.404565 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-encryption-config\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.410255 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-np6df" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.411377 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.421119 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-etcd-client\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.438753 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.451134 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.459710 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-image-import-ca\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.471591 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.471754 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/496e7e53-b6f6-48b1-9657-a091d21c301f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.471785 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5b28caec-1680-4b2f-a530-644790ea6d7e-srv-cert\") pod \"olm-operator-6b444d44fb-f7tfz\" (UID: \"5b28caec-1680-4b2f-a530-644790ea6d7e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.471820 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c421f8b0-bb08-4081-8071-9a5f7546630f-config\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.471844 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a8d300c-d03d-4927-b9ba-c17f692a9db4-config\") pod \"route-controller-manager-6576b87f9c-2gcb8\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.471866 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-config-volume\") pod \"collect-profiles-29322240-5zpj4\" (UID: \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.471885 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7ab8a342-ec8c-4466-bfaa-d86e384736d8-certs\") pod \"machine-config-server-6p58q\" (UID: \"7ab8a342-ec8c-4466-bfaa-d86e384736d8\") " pod="openshift-machine-config-operator/machine-config-server-6p58q" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.471922 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6k29\" (UniqueName: \"kubernetes.io/projected/e500be09-9f04-49a8-8ad2-e69012bd36f1-kube-api-access-d6k29\") pod \"multus-admission-controller-857f4d67dd-6w44d\" (UID: \"e500be09-9f04-49a8-8ad2-e69012bd36f1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6w44d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.471945 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts2ck\" (UniqueName: \"kubernetes.io/projected/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-kube-api-access-ts2ck\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.471967 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/939458a3-4eaf-46eb-aeb5-be7daf4f0f8c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bnh2z\" (UID: \"939458a3-4eaf-46eb-aeb5-be7daf4f0f8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.471993 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6zb9\" (UniqueName: \"kubernetes.io/projected/a44f6721-0fe0-45cc-8290-fe52ffb3f8f2-kube-api-access-b6zb9\") pod \"openshift-config-operator-7777fb866f-cg79v\" (UID: \"a44f6721-0fe0-45cc-8290-fe52ffb3f8f2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472016 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5450e1a6-a528-4242-8d81-a3241131dcc5-serving-cert\") pod \"service-ca-operator-777779d784-kjhjx\" (UID: \"5450e1a6-a528-4242-8d81-a3241131dcc5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472069 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f188f7e-b855-4902-b172-db25f2a53746-service-ca-bundle\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472108 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhm97\" (UniqueName: \"kubernetes.io/projected/10a86730-3fae-4053-b4e0-48effedbf3de-kube-api-access-rhm97\") pod \"machine-config-operator-74547568cd-xh5hv\" (UID: \"10a86730-3fae-4053-b4e0-48effedbf3de\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472144 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/496e7e53-b6f6-48b1-9657-a091d21c301f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472167 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c421f8b0-bb08-4081-8071-9a5f7546630f-serving-cert\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472201 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4739e7eb-ea30-427f-a701-29eb99ddd136-metrics-tls\") pod \"dns-default-l6d7x\" (UID: \"4739e7eb-ea30-427f-a701-29eb99ddd136\") " pod="openshift-dns/dns-default-l6d7x" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472240 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f371d8f0-5386-4a69-b81c-8c656d71cd54-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-54rxk\" (UID: \"f371d8f0-5386-4a69-b81c-8c656d71cd54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472264 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz2rc\" (UniqueName: \"kubernetes.io/projected/3f188f7e-b855-4902-b172-db25f2a53746-kube-api-access-xz2rc\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472286 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-service-ca\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472335 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/c421f8b0-bb08-4081-8071-9a5f7546630f-etcd-service-ca\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472357 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dg6w\" (UniqueName: \"kubernetes.io/projected/7ab8a342-ec8c-4466-bfaa-d86e384736d8-kube-api-access-8dg6w\") pod \"machine-config-server-6p58q\" (UID: \"7ab8a342-ec8c-4466-bfaa-d86e384736d8\") " pod="openshift-machine-config-operator/machine-config-server-6p58q" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472377 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4739e7eb-ea30-427f-a701-29eb99ddd136-config-volume\") pod \"dns-default-l6d7x\" (UID: \"4739e7eb-ea30-427f-a701-29eb99ddd136\") " pod="openshift-dns/dns-default-l6d7x" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472399 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/afecdb5b-56b5-4112-9e71-5f05e43302f0-srv-cert\") pod \"catalog-operator-68c6474976-w6hhh\" (UID: \"afecdb5b-56b5-4112-9e71-5f05e43302f0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472419 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-socket-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472453 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jgvx\" (UniqueName: \"kubernetes.io/projected/6d58f9c2-068f-4592-8de6-5217d0f51742-kube-api-access-2jgvx\") pod \"kube-storage-version-migrator-operator-b67b599dd-ns7lp\" (UID: \"6d58f9c2-068f-4592-8de6-5217d0f51742\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472490 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8db32339-b3a2-4b00-9352-575da2340e03-webhook-cert\") pod \"packageserver-d55dfcdfc-6sm6z\" (UID: \"8db32339-b3a2-4b00-9352-575da2340e03\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472512 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/afecdb5b-56b5-4112-9e71-5f05e43302f0-profile-collector-cert\") pod \"catalog-operator-68c6474976-w6hhh\" (UID: \"afecdb5b-56b5-4112-9e71-5f05e43302f0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472533 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-csi-data-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472568 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4bds\" (UniqueName: \"kubernetes.io/projected/e7f82b60-0afe-442d-ab72-40411e685fbc-kube-api-access-m4bds\") pod \"ingress-canary-5jsl4\" (UID: \"e7f82b60-0afe-442d-ab72-40411e685fbc\") " pod="openshift-ingress-canary/ingress-canary-5jsl4" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472602 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/496e7e53-b6f6-48b1-9657-a091d21c301f-registry-certificates\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472623 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/052caf82-2622-48b3-8c66-02955a93dcf5-config\") pod \"machine-approver-56656f9798-m4w6w\" (UID: \"052caf82-2622-48b3-8c66-02955a93dcf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472643 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trjk7\" (UniqueName: \"kubernetes.io/projected/8db32339-b3a2-4b00-9352-575da2340e03-kube-api-access-trjk7\") pod \"packageserver-d55dfcdfc-6sm6z\" (UID: \"8db32339-b3a2-4b00-9352-575da2340e03\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472667 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/71c3e368-7f4a-43e1-94a6-81374ba2fa00-signing-key\") pod \"service-ca-9c57cc56f-4r5zz\" (UID: \"71c3e368-7f4a-43e1-94a6-81374ba2fa00\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r5zz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472702 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/60142257-8641-4fba-83ca-bb1af9f68c0d-default-certificate\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472724 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-oauth-serving-cert\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472751 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98ea4e5f-eaa2-4612-81a5-5064a6dbc957-config\") pod \"kube-apiserver-operator-766d6c64bb-p7r2h\" (UID: \"98ea4e5f-eaa2-4612-81a5-5064a6dbc957\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472776 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82hfx\" (UniqueName: \"kubernetes.io/projected/a79834ba-7c32-4c45-a927-66cc2ebf3e97-kube-api-access-82hfx\") pod \"cluster-samples-operator-665b6dd947-k6p79\" (UID: \"a79834ba-7c32-4c45-a927-66cc2ebf3e97\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472792 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/052caf82-2622-48b3-8c66-02955a93dcf5-auth-proxy-config\") pod \"machine-approver-56656f9798-m4w6w\" (UID: \"052caf82-2622-48b3-8c66-02955a93dcf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472806 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-console-config\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472824 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5450e1a6-a528-4242-8d81-a3241131dcc5-config\") pod \"service-ca-operator-777779d784-kjhjx\" (UID: \"5450e1a6-a528-4242-8d81-a3241131dcc5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472857 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7smz\" (UniqueName: \"kubernetes.io/projected/052caf82-2622-48b3-8c66-02955a93dcf5-kube-api-access-d7smz\") pod \"machine-approver-56656f9798-m4w6w\" (UID: \"052caf82-2622-48b3-8c66-02955a93dcf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472881 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60142257-8641-4fba-83ca-bb1af9f68c0d-metrics-certs\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472897 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7n5q\" (UniqueName: \"kubernetes.io/projected/c421f8b0-bb08-4081-8071-9a5f7546630f-kube-api-access-l7n5q\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472922 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-bound-sa-token\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472943 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f9kw\" (UniqueName: \"kubernetes.io/projected/960e1b01-454c-4563-ba02-c35754a0b970-kube-api-access-7f9kw\") pod \"package-server-manager-789f6589d5-szpqh\" (UID: \"960e1b01-454c-4563-ba02-c35754a0b970\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472972 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98ea4e5f-eaa2-4612-81a5-5064a6dbc957-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-p7r2h\" (UID: \"98ea4e5f-eaa2-4612-81a5-5064a6dbc957\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.472997 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15670655-7e86-4731-bb0b-7ebd96a20e59-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fkxmw\" (UID: \"15670655-7e86-4731-bb0b-7ebd96a20e59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473020 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e500be09-9f04-49a8-8ad2-e69012bd36f1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6w44d\" (UID: \"e500be09-9f04-49a8-8ad2-e69012bd36f1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6w44d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473043 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/60142257-8641-4fba-83ca-bb1af9f68c0d-stats-auth\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473093 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a8d300c-d03d-4927-b9ba-c17f692a9db4-client-ca\") pod \"route-controller-manager-6576b87f9c-2gcb8\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473115 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqfmk\" (UniqueName: \"kubernetes.io/projected/5b28caec-1680-4b2f-a530-644790ea6d7e-kube-api-access-kqfmk\") pod \"olm-operator-6b444d44fb-f7tfz\" (UID: \"5b28caec-1680-4b2f-a530-644790ea6d7e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473139 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-registration-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473160 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/939458a3-4eaf-46eb-aeb5-be7daf4f0f8c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bnh2z\" (UID: \"939458a3-4eaf-46eb-aeb5-be7daf4f0f8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473198 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-registry-tls\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473221 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10a86730-3fae-4053-b4e0-48effedbf3de-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xh5hv\" (UID: \"10a86730-3fae-4053-b4e0-48effedbf3de\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473244 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-secret-volume\") pod \"collect-profiles-29322240-5zpj4\" (UID: \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473270 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwzxw\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-kube-api-access-pwzxw\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473290 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f188f7e-b855-4902-b172-db25f2a53746-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473347 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgmc4\" (UniqueName: \"kubernetes.io/projected/f371d8f0-5386-4a69-b81c-8c656d71cd54-kube-api-access-wgmc4\") pod \"machine-config-controller-84d6567774-54rxk\" (UID: \"f371d8f0-5386-4a69-b81c-8c656d71cd54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473395 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f371d8f0-5386-4a69-b81c-8c656d71cd54-proxy-tls\") pod \"machine-config-controller-84d6567774-54rxk\" (UID: \"f371d8f0-5386-4a69-b81c-8c656d71cd54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473432 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5zsg\" (UniqueName: \"kubernetes.io/projected/027579b0-86e0-4143-a1ad-8db0c669b7a2-kube-api-access-n5zsg\") pod \"migrator-59844c95c7-7rq8t\" (UID: \"027579b0-86e0-4143-a1ad-8db0c669b7a2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rq8t" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473456 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c442dae3-d66d-47a2-b1dc-30332e947203-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9qvwb\" (UID: \"c442dae3-d66d-47a2-b1dc-30332e947203\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473479 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx8qg\" (UniqueName: \"kubernetes.io/projected/c442dae3-d66d-47a2-b1dc-30332e947203-kube-api-access-dx8qg\") pod \"marketplace-operator-79b997595-9qvwb\" (UID: \"c442dae3-d66d-47a2-b1dc-30332e947203\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473502 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a8d300c-d03d-4927-b9ba-c17f692a9db4-serving-cert\") pod \"route-controller-manager-6576b87f9c-2gcb8\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473521 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8db32339-b3a2-4b00-9352-575da2340e03-tmpfs\") pod \"packageserver-d55dfcdfc-6sm6z\" (UID: \"8db32339-b3a2-4b00-9352-575da2340e03\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473550 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10a86730-3fae-4053-b4e0-48effedbf3de-proxy-tls\") pod \"machine-config-operator-74547568cd-xh5hv\" (UID: \"10a86730-3fae-4053-b4e0-48effedbf3de\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473572 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a44f6721-0fe0-45cc-8290-fe52ffb3f8f2-serving-cert\") pod \"openshift-config-operator-7777fb866f-cg79v\" (UID: \"a44f6721-0fe0-45cc-8290-fe52ffb3f8f2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473594 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c421f8b0-bb08-4081-8071-9a5f7546630f-etcd-client\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473642 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcp85\" (UniqueName: \"kubernetes.io/projected/8282f47b-cd59-4cba-a1aa-7528d57842ec-kube-api-access-qcp85\") pod \"control-plane-machine-set-operator-78cbb6b69f-z4ml2\" (UID: \"8282f47b-cd59-4cba-a1aa-7528d57842ec\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z4ml2" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473681 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dns5\" (UniqueName: \"kubernetes.io/projected/b8cf6941-f833-4465-a3e9-9efea4b9b5b0-kube-api-access-7dns5\") pod \"downloads-7954f5f757-5tjmt\" (UID: \"b8cf6941-f833-4465-a3e9-9efea4b9b5b0\") " pod="openshift-console/downloads-7954f5f757-5tjmt" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473729 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/496e7e53-b6f6-48b1-9657-a091d21c301f-trusted-ca\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473763 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-plugins-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473786 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tqlj\" (UniqueName: \"kubernetes.io/projected/71c3e368-7f4a-43e1-94a6-81374ba2fa00-kube-api-access-8tqlj\") pod \"service-ca-9c57cc56f-4r5zz\" (UID: \"71c3e368-7f4a-43e1-94a6-81374ba2fa00\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r5zz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473847 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f188f7e-b855-4902-b172-db25f2a53746-config\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473872 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgpvw\" (UniqueName: \"kubernetes.io/projected/60142257-8641-4fba-83ca-bb1af9f68c0d-kube-api-access-tgpvw\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473895 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a79834ba-7c32-4c45-a927-66cc2ebf3e97-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k6p79\" (UID: \"a79834ba-7c32-4c45-a927-66cc2ebf3e97\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473920 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqkk7\" (UniqueName: \"kubernetes.io/projected/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-kube-api-access-mqkk7\") pod \"collect-profiles-29322240-5zpj4\" (UID: \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473944 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/79d896b5-b34d-4206-92d8-99caeafdffb2-console-serving-cert\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.473970 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmj5x\" (UniqueName: \"kubernetes.io/projected/5450e1a6-a528-4242-8d81-a3241131dcc5-kube-api-access-vmj5x\") pod \"service-ca-operator-777779d784-kjhjx\" (UID: \"5450e1a6-a528-4242-8d81-a3241131dcc5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.474012 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/939458a3-4eaf-46eb-aeb5-be7daf4f0f8c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bnh2z\" (UID: \"939458a3-4eaf-46eb-aeb5-be7daf4f0f8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.474038 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-trusted-ca-bundle\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.474133 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 01 16:01:36 crc kubenswrapper[4726]: E1001 16:01:36.474372 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:36.974353068 +0000 UTC m=+149.875905655 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.476610 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a8d300c-d03d-4927-b9ba-c17f692a9db4-client-ca\") pod \"route-controller-manager-6576b87f9c-2gcb8\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.476782 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a8d300c-d03d-4927-b9ba-c17f692a9db4-config\") pod \"route-controller-manager-6576b87f9c-2gcb8\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.477364 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/496e7e53-b6f6-48b1-9657-a091d21c301f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.477393 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15670655-7e86-4731-bb0b-7ebd96a20e59-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fkxmw\" (UID: \"15670655-7e86-4731-bb0b-7ebd96a20e59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.478266 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f371d8f0-5386-4a69-b81c-8c656d71cd54-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-54rxk\" (UID: \"f371d8f0-5386-4a69-b81c-8c656d71cd54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.478915 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-service-ca\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.479385 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10a86730-3fae-4053-b4e0-48effedbf3de-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xh5hv\" (UID: \"10a86730-3fae-4053-b4e0-48effedbf3de\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.479641 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f188f7e-b855-4902-b172-db25f2a53746-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.479803 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f188f7e-b855-4902-b172-db25f2a53746-config\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.481340 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/496e7e53-b6f6-48b1-9657-a091d21c301f-trusted-ca\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.483356 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt7mt\" (UniqueName: \"kubernetes.io/projected/1a8d300c-d03d-4927-b9ba-c17f692a9db4-kube-api-access-nt7mt\") pod \"route-controller-manager-6576b87f9c-2gcb8\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.483935 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5b28caec-1680-4b2f-a530-644790ea6d7e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-f7tfz\" (UID: \"5b28caec-1680-4b2f-a530-644790ea6d7e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.483986 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f188f7e-b855-4902-b172-db25f2a53746-serving-cert\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.484125 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15670655-7e86-4731-bb0b-7ebd96a20e59-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fkxmw\" (UID: \"15670655-7e86-4731-bb0b-7ebd96a20e59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.484239 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7tsz\" (UniqueName: \"kubernetes.io/projected/15670655-7e86-4731-bb0b-7ebd96a20e59-kube-api-access-g7tsz\") pod \"openshift-controller-manager-operator-756b6f6bc6-fkxmw\" (UID: \"15670655-7e86-4731-bb0b-7ebd96a20e59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.484253 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e500be09-9f04-49a8-8ad2-e69012bd36f1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6w44d\" (UID: \"e500be09-9f04-49a8-8ad2-e69012bd36f1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6w44d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.485650 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/052caf82-2622-48b3-8c66-02955a93dcf5-config\") pod \"machine-approver-56656f9798-m4w6w\" (UID: \"052caf82-2622-48b3-8c66-02955a93dcf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.486073 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/052caf82-2622-48b3-8c66-02955a93dcf5-auth-proxy-config\") pod \"machine-approver-56656f9798-m4w6w\" (UID: \"052caf82-2622-48b3-8c66-02955a93dcf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.486423 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/496e7e53-b6f6-48b1-9657-a091d21c301f-registry-certificates\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.486472 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f188f7e-b855-4902-b172-db25f2a53746-service-ca-bundle\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.486542 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr"] Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.486624 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8282f47b-cd59-4cba-a1aa-7528d57842ec-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z4ml2\" (UID: \"8282f47b-cd59-4cba-a1aa-7528d57842ec\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z4ml2" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.486670 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/10a86730-3fae-4053-b4e0-48effedbf3de-images\") pod \"machine-config-operator-74547568cd-xh5hv\" (UID: \"10a86730-3fae-4053-b4e0-48effedbf3de\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.486695 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c442dae3-d66d-47a2-b1dc-30332e947203-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9qvwb\" (UID: \"c442dae3-d66d-47a2-b1dc-30332e947203\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.486713 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-mountpoint-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.486733 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60142257-8641-4fba-83ca-bb1af9f68c0d-service-ca-bundle\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.486753 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/c421f8b0-bb08-4081-8071-9a5f7546630f-etcd-ca\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.487022 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-oauth-serving-cert\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.487039 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zl4l\" (UniqueName: \"kubernetes.io/projected/afecdb5b-56b5-4112-9e71-5f05e43302f0-kube-api-access-5zl4l\") pod \"catalog-operator-68c6474976-w6hhh\" (UID: \"afecdb5b-56b5-4112-9e71-5f05e43302f0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.487281 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/10a86730-3fae-4053-b4e0-48effedbf3de-images\") pod \"machine-config-operator-74547568cd-xh5hv\" (UID: \"10a86730-3fae-4053-b4e0-48effedbf3de\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.487498 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8db32339-b3a2-4b00-9352-575da2340e03-apiservice-cert\") pod \"packageserver-d55dfcdfc-6sm6z\" (UID: \"8db32339-b3a2-4b00-9352-575da2340e03\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.487528 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7f82b60-0afe-442d-ab72-40411e685fbc-cert\") pod \"ingress-canary-5jsl4\" (UID: \"e7f82b60-0afe-442d-ab72-40411e685fbc\") " pod="openshift-ingress-canary/ingress-canary-5jsl4" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.487576 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d58f9c2-068f-4592-8de6-5217d0f51742-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ns7lp\" (UID: \"6d58f9c2-068f-4592-8de6-5217d0f51742\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.487604 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/052caf82-2622-48b3-8c66-02955a93dcf5-machine-approver-tls\") pod \"machine-approver-56656f9798-m4w6w\" (UID: \"052caf82-2622-48b3-8c66-02955a93dcf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.487631 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98ea4e5f-eaa2-4612-81a5-5064a6dbc957-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-p7r2h\" (UID: \"98ea4e5f-eaa2-4612-81a5-5064a6dbc957\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.487659 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/960e1b01-454c-4563-ba02-c35754a0b970-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-szpqh\" (UID: \"960e1b01-454c-4563-ba02-c35754a0b970\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.487886 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a44f6721-0fe0-45cc-8290-fe52ffb3f8f2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cg79v\" (UID: \"a44f6721-0fe0-45cc-8290-fe52ffb3f8f2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.487949 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7ab8a342-ec8c-4466-bfaa-d86e384736d8-node-bootstrap-token\") pod \"machine-config-server-6p58q\" (UID: \"7ab8a342-ec8c-4466-bfaa-d86e384736d8\") " pod="openshift-machine-config-operator/machine-config-server-6p58q" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.487972 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/79d896b5-b34d-4206-92d8-99caeafdffb2-console-oauth-config\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.487989 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqg5p\" (UniqueName: \"kubernetes.io/projected/4739e7eb-ea30-427f-a701-29eb99ddd136-kube-api-access-nqg5p\") pod \"dns-default-l6d7x\" (UID: \"4739e7eb-ea30-427f-a701-29eb99ddd136\") " pod="openshift-dns/dns-default-l6d7x" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.488299 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a44f6721-0fe0-45cc-8290-fe52ffb3f8f2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cg79v\" (UID: \"a44f6721-0fe0-45cc-8290-fe52ffb3f8f2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.488425 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60142257-8641-4fba-83ca-bb1af9f68c0d-service-ca-bundle\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.488738 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x82vr\" (UniqueName: \"kubernetes.io/projected/79d896b5-b34d-4206-92d8-99caeafdffb2-kube-api-access-x82vr\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.490619 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/71c3e368-7f4a-43e1-94a6-81374ba2fa00-signing-cabundle\") pod \"service-ca-9c57cc56f-4r5zz\" (UID: \"71c3e368-7f4a-43e1-94a6-81374ba2fa00\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r5zz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.490649 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d58f9c2-068f-4592-8de6-5217d0f51742-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ns7lp\" (UID: \"6d58f9c2-068f-4592-8de6-5217d0f51742\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.489276 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-trusted-ca-bundle\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.490773 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a79834ba-7c32-4c45-a927-66cc2ebf3e97-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k6p79\" (UID: \"a79834ba-7c32-4c45-a927-66cc2ebf3e97\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.491544 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-console-config\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.491765 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.494220 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f371d8f0-5386-4a69-b81c-8c656d71cd54-proxy-tls\") pod \"machine-config-controller-84d6567774-54rxk\" (UID: \"f371d8f0-5386-4a69-b81c-8c656d71cd54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.496377 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/60142257-8641-4fba-83ca-bb1af9f68c0d-stats-auth\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.496395 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/496e7e53-b6f6-48b1-9657-a091d21c301f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.496417 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a8d300c-d03d-4927-b9ba-c17f692a9db4-serving-cert\") pod \"route-controller-manager-6576b87f9c-2gcb8\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.496656 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10a86730-3fae-4053-b4e0-48effedbf3de-proxy-tls\") pod \"machine-config-operator-74547568cd-xh5hv\" (UID: \"10a86730-3fae-4053-b4e0-48effedbf3de\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.497881 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a44f6721-0fe0-45cc-8290-fe52ffb3f8f2-serving-cert\") pod \"openshift-config-operator-7777fb866f-cg79v\" (UID: \"a44f6721-0fe0-45cc-8290-fe52ffb3f8f2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.498120 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60142257-8641-4fba-83ca-bb1af9f68c0d-metrics-certs\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.498517 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15670655-7e86-4731-bb0b-7ebd96a20e59-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fkxmw\" (UID: \"15670655-7e86-4731-bb0b-7ebd96a20e59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.498560 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/60142257-8641-4fba-83ca-bb1af9f68c0d-default-certificate\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.499020 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-etcd-serving-ca\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.499965 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-registry-tls\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.500028 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/79d896b5-b34d-4206-92d8-99caeafdffb2-console-serving-cert\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.500586 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8282f47b-cd59-4cba-a1aa-7528d57842ec-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z4ml2\" (UID: \"8282f47b-cd59-4cba-a1aa-7528d57842ec\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z4ml2" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.500925 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f188f7e-b855-4902-b172-db25f2a53746-serving-cert\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.501509 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/79d896b5-b34d-4206-92d8-99caeafdffb2-console-oauth-config\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.501797 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/052caf82-2622-48b3-8c66-02955a93dcf5-machine-approver-tls\") pod \"machine-approver-56656f9798-m4w6w\" (UID: \"052caf82-2622-48b3-8c66-02955a93dcf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.518614 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.520382 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.538474 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.549648 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5399581d-2111-47d1-bfe7-007b7c46d706-images\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.550987 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.555321 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr" event={"ID":"7f189676-cdd8-4138-b73f-863e1c075bc2","Type":"ContainerStarted","Data":"825c6721dfc739350d08b8bd73364da0df212d96082a8b7ea05bba8cd0485efd"} Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.558398 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" event={"ID":"262d42bb-7bb1-49f1-8cf0-ad225e6dd673","Type":"ContainerStarted","Data":"59b7561c68cdede1e13fc76b9946aa3f755b1307b4cf067bd2ce7ac7b6c08af8"} Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.570671 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.590642 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591281 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqkk7\" (UniqueName: \"kubernetes.io/projected/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-kube-api-access-mqkk7\") pod \"collect-profiles-29322240-5zpj4\" (UID: \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591316 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmj5x\" (UniqueName: \"kubernetes.io/projected/5450e1a6-a528-4242-8d81-a3241131dcc5-kube-api-access-vmj5x\") pod \"service-ca-operator-777779d784-kjhjx\" (UID: \"5450e1a6-a528-4242-8d81-a3241131dcc5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591342 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/939458a3-4eaf-46eb-aeb5-be7daf4f0f8c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bnh2z\" (UID: \"939458a3-4eaf-46eb-aeb5-be7daf4f0f8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591380 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5b28caec-1680-4b2f-a530-644790ea6d7e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-f7tfz\" (UID: \"5b28caec-1680-4b2f-a530-644790ea6d7e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591445 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-mountpoint-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591461 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c442dae3-d66d-47a2-b1dc-30332e947203-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9qvwb\" (UID: \"c442dae3-d66d-47a2-b1dc-30332e947203\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591478 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/c421f8b0-bb08-4081-8071-9a5f7546630f-etcd-ca\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591508 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zl4l\" (UniqueName: \"kubernetes.io/projected/afecdb5b-56b5-4112-9e71-5f05e43302f0-kube-api-access-5zl4l\") pod \"catalog-operator-68c6474976-w6hhh\" (UID: \"afecdb5b-56b5-4112-9e71-5f05e43302f0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591536 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7f82b60-0afe-442d-ab72-40411e685fbc-cert\") pod \"ingress-canary-5jsl4\" (UID: \"e7f82b60-0afe-442d-ab72-40411e685fbc\") " pod="openshift-ingress-canary/ingress-canary-5jsl4" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591555 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8db32339-b3a2-4b00-9352-575da2340e03-apiservice-cert\") pod \"packageserver-d55dfcdfc-6sm6z\" (UID: \"8db32339-b3a2-4b00-9352-575da2340e03\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591571 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d58f9c2-068f-4592-8de6-5217d0f51742-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ns7lp\" (UID: \"6d58f9c2-068f-4592-8de6-5217d0f51742\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591596 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98ea4e5f-eaa2-4612-81a5-5064a6dbc957-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-p7r2h\" (UID: \"98ea4e5f-eaa2-4612-81a5-5064a6dbc957\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591628 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7ab8a342-ec8c-4466-bfaa-d86e384736d8-node-bootstrap-token\") pod \"machine-config-server-6p58q\" (UID: \"7ab8a342-ec8c-4466-bfaa-d86e384736d8\") " pod="openshift-machine-config-operator/machine-config-server-6p58q" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591655 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/960e1b01-454c-4563-ba02-c35754a0b970-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-szpqh\" (UID: \"960e1b01-454c-4563-ba02-c35754a0b970\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591682 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqg5p\" (UniqueName: \"kubernetes.io/projected/4739e7eb-ea30-427f-a701-29eb99ddd136-kube-api-access-nqg5p\") pod \"dns-default-l6d7x\" (UID: \"4739e7eb-ea30-427f-a701-29eb99ddd136\") " pod="openshift-dns/dns-default-l6d7x" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591704 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d58f9c2-068f-4592-8de6-5217d0f51742-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ns7lp\" (UID: \"6d58f9c2-068f-4592-8de6-5217d0f51742\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591743 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/71c3e368-7f4a-43e1-94a6-81374ba2fa00-signing-cabundle\") pod \"service-ca-9c57cc56f-4r5zz\" (UID: \"71c3e368-7f4a-43e1-94a6-81374ba2fa00\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r5zz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591764 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5b28caec-1680-4b2f-a530-644790ea6d7e-srv-cert\") pod \"olm-operator-6b444d44fb-f7tfz\" (UID: \"5b28caec-1680-4b2f-a530-644790ea6d7e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591793 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c421f8b0-bb08-4081-8071-9a5f7546630f-config\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591813 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-config-volume\") pod \"collect-profiles-29322240-5zpj4\" (UID: \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591831 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7ab8a342-ec8c-4466-bfaa-d86e384736d8-certs\") pod \"machine-config-server-6p58q\" (UID: \"7ab8a342-ec8c-4466-bfaa-d86e384736d8\") " pod="openshift-machine-config-operator/machine-config-server-6p58q" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591859 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts2ck\" (UniqueName: \"kubernetes.io/projected/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-kube-api-access-ts2ck\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591885 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/939458a3-4eaf-46eb-aeb5-be7daf4f0f8c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bnh2z\" (UID: \"939458a3-4eaf-46eb-aeb5-be7daf4f0f8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591905 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5450e1a6-a528-4242-8d81-a3241131dcc5-serving-cert\") pod \"service-ca-operator-777779d784-kjhjx\" (UID: \"5450e1a6-a528-4242-8d81-a3241131dcc5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591937 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c421f8b0-bb08-4081-8071-9a5f7546630f-serving-cert\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591960 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4739e7eb-ea30-427f-a701-29eb99ddd136-metrics-tls\") pod \"dns-default-l6d7x\" (UID: \"4739e7eb-ea30-427f-a701-29eb99ddd136\") " pod="openshift-dns/dns-default-l6d7x" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.591995 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/c421f8b0-bb08-4081-8071-9a5f7546630f-etcd-service-ca\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592018 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dg6w\" (UniqueName: \"kubernetes.io/projected/7ab8a342-ec8c-4466-bfaa-d86e384736d8-kube-api-access-8dg6w\") pod \"machine-config-server-6p58q\" (UID: \"7ab8a342-ec8c-4466-bfaa-d86e384736d8\") " pod="openshift-machine-config-operator/machine-config-server-6p58q" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592040 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-socket-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592081 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4739e7eb-ea30-427f-a701-29eb99ddd136-config-volume\") pod \"dns-default-l6d7x\" (UID: \"4739e7eb-ea30-427f-a701-29eb99ddd136\") " pod="openshift-dns/dns-default-l6d7x" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592102 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/afecdb5b-56b5-4112-9e71-5f05e43302f0-srv-cert\") pod \"catalog-operator-68c6474976-w6hhh\" (UID: \"afecdb5b-56b5-4112-9e71-5f05e43302f0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592120 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jgvx\" (UniqueName: \"kubernetes.io/projected/6d58f9c2-068f-4592-8de6-5217d0f51742-kube-api-access-2jgvx\") pod \"kube-storage-version-migrator-operator-b67b599dd-ns7lp\" (UID: \"6d58f9c2-068f-4592-8de6-5217d0f51742\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592138 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8db32339-b3a2-4b00-9352-575da2340e03-webhook-cert\") pod \"packageserver-d55dfcdfc-6sm6z\" (UID: \"8db32339-b3a2-4b00-9352-575da2340e03\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592155 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-csi-data-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592172 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/afecdb5b-56b5-4112-9e71-5f05e43302f0-profile-collector-cert\") pod \"catalog-operator-68c6474976-w6hhh\" (UID: \"afecdb5b-56b5-4112-9e71-5f05e43302f0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592189 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4bds\" (UniqueName: \"kubernetes.io/projected/e7f82b60-0afe-442d-ab72-40411e685fbc-kube-api-access-m4bds\") pod \"ingress-canary-5jsl4\" (UID: \"e7f82b60-0afe-442d-ab72-40411e685fbc\") " pod="openshift-ingress-canary/ingress-canary-5jsl4" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592195 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/939458a3-4eaf-46eb-aeb5-be7daf4f0f8c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bnh2z\" (UID: \"939458a3-4eaf-46eb-aeb5-be7daf4f0f8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592206 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trjk7\" (UniqueName: \"kubernetes.io/projected/8db32339-b3a2-4b00-9352-575da2340e03-kube-api-access-trjk7\") pod \"packageserver-d55dfcdfc-6sm6z\" (UID: \"8db32339-b3a2-4b00-9352-575da2340e03\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592269 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/71c3e368-7f4a-43e1-94a6-81374ba2fa00-signing-key\") pod \"service-ca-9c57cc56f-4r5zz\" (UID: \"71c3e368-7f4a-43e1-94a6-81374ba2fa00\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r5zz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592313 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98ea4e5f-eaa2-4612-81a5-5064a6dbc957-config\") pod \"kube-apiserver-operator-766d6c64bb-p7r2h\" (UID: \"98ea4e5f-eaa2-4612-81a5-5064a6dbc957\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592343 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5450e1a6-a528-4242-8d81-a3241131dcc5-config\") pod \"service-ca-operator-777779d784-kjhjx\" (UID: \"5450e1a6-a528-4242-8d81-a3241131dcc5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592390 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7n5q\" (UniqueName: \"kubernetes.io/projected/c421f8b0-bb08-4081-8071-9a5f7546630f-kube-api-access-l7n5q\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592433 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f9kw\" (UniqueName: \"kubernetes.io/projected/960e1b01-454c-4563-ba02-c35754a0b970-kube-api-access-7f9kw\") pod \"package-server-manager-789f6589d5-szpqh\" (UID: \"960e1b01-454c-4563-ba02-c35754a0b970\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592462 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98ea4e5f-eaa2-4612-81a5-5064a6dbc957-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-p7r2h\" (UID: \"98ea4e5f-eaa2-4612-81a5-5064a6dbc957\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592488 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqfmk\" (UniqueName: \"kubernetes.io/projected/5b28caec-1680-4b2f-a530-644790ea6d7e-kube-api-access-kqfmk\") pod \"olm-operator-6b444d44fb-f7tfz\" (UID: \"5b28caec-1680-4b2f-a530-644790ea6d7e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592493 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c442dae3-d66d-47a2-b1dc-30332e947203-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9qvwb\" (UID: \"c442dae3-d66d-47a2-b1dc-30332e947203\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592510 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-registration-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592532 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/939458a3-4eaf-46eb-aeb5-be7daf4f0f8c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bnh2z\" (UID: \"939458a3-4eaf-46eb-aeb5-be7daf4f0f8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592557 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-secret-volume\") pod \"collect-profiles-29322240-5zpj4\" (UID: \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592589 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592656 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c442dae3-d66d-47a2-b1dc-30332e947203-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9qvwb\" (UID: \"c442dae3-d66d-47a2-b1dc-30332e947203\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592681 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx8qg\" (UniqueName: \"kubernetes.io/projected/c442dae3-d66d-47a2-b1dc-30332e947203-kube-api-access-dx8qg\") pod \"marketplace-operator-79b997595-9qvwb\" (UID: \"c442dae3-d66d-47a2-b1dc-30332e947203\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592708 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5zsg\" (UniqueName: \"kubernetes.io/projected/027579b0-86e0-4143-a1ad-8db0c669b7a2-kube-api-access-n5zsg\") pod \"migrator-59844c95c7-7rq8t\" (UID: \"027579b0-86e0-4143-a1ad-8db0c669b7a2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rq8t" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592731 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8db32339-b3a2-4b00-9352-575da2340e03-tmpfs\") pod \"packageserver-d55dfcdfc-6sm6z\" (UID: \"8db32339-b3a2-4b00-9352-575da2340e03\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592768 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c421f8b0-bb08-4081-8071-9a5f7546630f-etcd-client\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592813 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-plugins-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592837 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tqlj\" (UniqueName: \"kubernetes.io/projected/71c3e368-7f4a-43e1-94a6-81374ba2fa00-kube-api-access-8tqlj\") pod \"service-ca-9c57cc56f-4r5zz\" (UID: \"71c3e368-7f4a-43e1-94a6-81374ba2fa00\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r5zz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592888 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d58f9c2-068f-4592-8de6-5217d0f51742-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ns7lp\" (UID: \"6d58f9c2-068f-4592-8de6-5217d0f51742\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.593269 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/c421f8b0-bb08-4081-8071-9a5f7546630f-etcd-ca\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: E1001 16:01:36.593702 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.093688003 +0000 UTC m=+149.995240580 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.594472 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8db32339-b3a2-4b00-9352-575da2340e03-tmpfs\") pod \"packageserver-d55dfcdfc-6sm6z\" (UID: \"8db32339-b3a2-4b00-9352-575da2340e03\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.594663 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-csi-data-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.592558 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-mountpoint-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.595478 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c421f8b0-bb08-4081-8071-9a5f7546630f-config\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.595836 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-socket-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.595997 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/c421f8b0-bb08-4081-8071-9a5f7546630f-etcd-service-ca\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.596773 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-registration-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.596867 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98ea4e5f-eaa2-4612-81a5-5064a6dbc957-config\") pod \"kube-apiserver-operator-766d6c64bb-p7r2h\" (UID: \"98ea4e5f-eaa2-4612-81a5-5064a6dbc957\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.597235 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/71c3e368-7f4a-43e1-94a6-81374ba2fa00-signing-cabundle\") pod \"service-ca-9c57cc56f-4r5zz\" (UID: \"71c3e368-7f4a-43e1-94a6-81374ba2fa00\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r5zz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.597296 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5b28caec-1680-4b2f-a530-644790ea6d7e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-f7tfz\" (UID: \"5b28caec-1680-4b2f-a530-644790ea6d7e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.597778 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8db32339-b3a2-4b00-9352-575da2340e03-apiservice-cert\") pod \"packageserver-d55dfcdfc-6sm6z\" (UID: \"8db32339-b3a2-4b00-9352-575da2340e03\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.597919 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4739e7eb-ea30-427f-a701-29eb99ddd136-config-volume\") pod \"dns-default-l6d7x\" (UID: \"4739e7eb-ea30-427f-a701-29eb99ddd136\") " pod="openshift-dns/dns-default-l6d7x" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.597944 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c421f8b0-bb08-4081-8071-9a5f7546630f-etcd-client\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.598502 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52fd2\" (UniqueName: \"kubernetes.io/projected/5399581d-2111-47d1-bfe7-007b7c46d706-kube-api-access-52fd2\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.598569 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5450e1a6-a528-4242-8d81-a3241131dcc5-config\") pod \"service-ca-operator-777779d784-kjhjx\" (UID: \"5450e1a6-a528-4242-8d81-a3241131dcc5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.598600 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d58f9c2-068f-4592-8de6-5217d0f51742-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ns7lp\" (UID: \"6d58f9c2-068f-4592-8de6-5217d0f51742\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.598681 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-plugins-dir\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.599031 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/71c3e368-7f4a-43e1-94a6-81374ba2fa00-signing-key\") pod \"service-ca-9c57cc56f-4r5zz\" (UID: \"71c3e368-7f4a-43e1-94a6-81374ba2fa00\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r5zz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.599818 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7f82b60-0afe-442d-ab72-40411e685fbc-cert\") pod \"ingress-canary-5jsl4\" (UID: \"e7f82b60-0afe-442d-ab72-40411e685fbc\") " pod="openshift-ingress-canary/ingress-canary-5jsl4" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.600345 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/afecdb5b-56b5-4112-9e71-5f05e43302f0-profile-collector-cert\") pod \"catalog-operator-68c6474976-w6hhh\" (UID: \"afecdb5b-56b5-4112-9e71-5f05e43302f0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.601285 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5450e1a6-a528-4242-8d81-a3241131dcc5-serving-cert\") pod \"service-ca-operator-777779d784-kjhjx\" (UID: \"5450e1a6-a528-4242-8d81-a3241131dcc5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.601412 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c442dae3-d66d-47a2-b1dc-30332e947203-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9qvwb\" (UID: \"c442dae3-d66d-47a2-b1dc-30332e947203\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.601519 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8db32339-b3a2-4b00-9352-575da2340e03-webhook-cert\") pod \"packageserver-d55dfcdfc-6sm6z\" (UID: \"8db32339-b3a2-4b00-9352-575da2340e03\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.601958 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5b28caec-1680-4b2f-a530-644790ea6d7e-srv-cert\") pod \"olm-operator-6b444d44fb-f7tfz\" (UID: \"5b28caec-1680-4b2f-a530-644790ea6d7e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.602354 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98ea4e5f-eaa2-4612-81a5-5064a6dbc957-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-p7r2h\" (UID: \"98ea4e5f-eaa2-4612-81a5-5064a6dbc957\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.602753 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7ab8a342-ec8c-4466-bfaa-d86e384736d8-node-bootstrap-token\") pod \"machine-config-server-6p58q\" (UID: \"7ab8a342-ec8c-4466-bfaa-d86e384736d8\") " pod="openshift-machine-config-operator/machine-config-server-6p58q" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.602859 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-config-volume\") pod \"collect-profiles-29322240-5zpj4\" (UID: \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.603179 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c421f8b0-bb08-4081-8071-9a5f7546630f-serving-cert\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.603813 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/960e1b01-454c-4563-ba02-c35754a0b970-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-szpqh\" (UID: \"960e1b01-454c-4563-ba02-c35754a0b970\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.604037 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4739e7eb-ea30-427f-a701-29eb99ddd136-metrics-tls\") pod \"dns-default-l6d7x\" (UID: \"4739e7eb-ea30-427f-a701-29eb99ddd136\") " pod="openshift-dns/dns-default-l6d7x" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.604117 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/afecdb5b-56b5-4112-9e71-5f05e43302f0-srv-cert\") pod \"catalog-operator-68c6474976-w6hhh\" (UID: \"afecdb5b-56b5-4112-9e71-5f05e43302f0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.604316 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-secret-volume\") pod \"collect-profiles-29322240-5zpj4\" (UID: \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.605404 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/939458a3-4eaf-46eb-aeb5-be7daf4f0f8c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bnh2z\" (UID: \"939458a3-4eaf-46eb-aeb5-be7daf4f0f8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.606173 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7ab8a342-ec8c-4466-bfaa-d86e384736d8-certs\") pod \"machine-config-server-6p58q\" (UID: \"7ab8a342-ec8c-4466-bfaa-d86e384736d8\") " pod="openshift-machine-config-operator/machine-config-server-6p58q" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.613177 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.620729 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5399581d-2111-47d1-bfe7-007b7c46d706-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.631643 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.642784 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-serving-cert\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.649666 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 01 16:01:36 crc kubenswrapper[4726]: E1001 16:01:36.659206 4726 configmap.go:193] Couldn't get configMap openshift-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:36 crc kubenswrapper[4726]: E1001 16:01:36.659272 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-audit podName:e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.659252608 +0000 UTC m=+150.560805185 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "audit" (UniqueName: "kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-audit") pod "apiserver-76f77b778f-vxgvr" (UID: "e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f") : failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:36 crc kubenswrapper[4726]: E1001 16:01:36.659564 4726 configmap.go:193] Couldn't get configMap openshift-apiserver/config: failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:36 crc kubenswrapper[4726]: E1001 16:01:36.659594 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-config podName:e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.659584247 +0000 UTC m=+150.561136824 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-config") pod "apiserver-76f77b778f-vxgvr" (UID: "e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f") : failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.664531 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5399581d-2111-47d1-bfe7-007b7c46d706-config\") pod \"machine-api-operator-5694c8668f-mpfpx\" (UID: \"5399581d-2111-47d1-bfe7-007b7c46d706\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.670135 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.670392 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-np6df"] Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.690885 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.693442 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:36 crc kubenswrapper[4726]: E1001 16:01:36.693593 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.193571021 +0000 UTC m=+150.095123598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.693984 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: E1001 16:01:36.694406 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.194395963 +0000 UTC m=+150.095948540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.701760 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" Oct 01 16:01:36 crc kubenswrapper[4726]: E1001 16:01:36.702698 4726 projected.go:288] Couldn't get configMap openshift-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:36 crc kubenswrapper[4726]: E1001 16:01:36.702741 4726 projected.go:194] Error preparing data for projected volume kube-api-access-jhwlw for pod openshift-apiserver/apiserver-76f77b778f-vxgvr: failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:36 crc kubenswrapper[4726]: E1001 16:01:36.702789 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-kube-api-access-jhwlw podName:e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.202771746 +0000 UTC m=+150.104324323 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-jhwlw" (UniqueName: "kubernetes.io/projected/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-kube-api-access-jhwlw") pod "apiserver-76f77b778f-vxgvr" (UID: "e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f") : failed to sync configmap cache: timed out waiting for the condition Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.709738 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.716345 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7jshc"] Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.731185 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zhfrh"] Oct 01 16:01:36 crc kubenswrapper[4726]: W1001 16:01:36.752342 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73789e0a_a83d_4470_9f4c_137b3c317b85.slice/crio-28894c50be0b5ef619576dd433216ee43b3d48b8cf5f6b9065b21a32fb43d9fb WatchSource:0}: Error finding container 28894c50be0b5ef619576dd433216ee43b3d48b8cf5f6b9065b21a32fb43d9fb: Status 404 returned error can't find the container with id 28894c50be0b5ef619576dd433216ee43b3d48b8cf5f6b9065b21a32fb43d9fb Oct 01 16:01:36 crc kubenswrapper[4726]: W1001 16:01:36.752586 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod654d70bb_a9aa_48d1_99dd_95a50b221440.slice/crio-0563a05056f5e7c7913ea4a1f0124476d0df838114ec618fc985717162ccabed WatchSource:0}: Error finding container 0563a05056f5e7c7913ea4a1f0124476d0df838114ec618fc985717162ccabed: Status 404 returned error can't find the container with id 0563a05056f5e7c7913ea4a1f0124476d0df838114ec618fc985717162ccabed Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.765758 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-bound-sa-token\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.788349 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7smz\" (UniqueName: \"kubernetes.io/projected/052caf82-2622-48b3-8c66-02955a93dcf5-kube-api-access-d7smz\") pod \"machine-approver-56656f9798-m4w6w\" (UID: \"052caf82-2622-48b3-8c66-02955a93dcf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.797604 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:36 crc kubenswrapper[4726]: E1001 16:01:36.798391 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.29837401 +0000 UTC m=+150.199926587 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.804163 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pq5pm"] Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.805392 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2"] Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.806865 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhm97\" (UniqueName: \"kubernetes.io/projected/10a86730-3fae-4053-b4e0-48effedbf3de-kube-api-access-rhm97\") pod \"machine-config-operator-74547568cd-xh5hv\" (UID: \"10a86730-3fae-4053-b4e0-48effedbf3de\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.826719 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz2rc\" (UniqueName: \"kubernetes.io/projected/3f188f7e-b855-4902-b172-db25f2a53746-kube-api-access-xz2rc\") pod \"authentication-operator-69f744f599-pxlbv\" (UID: \"3f188f7e-b855-4902-b172-db25f2a53746\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.849010 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwzxw\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-kube-api-access-pwzxw\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.851695 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558"] Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.863238 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx"] Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.865483 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgmc4\" (UniqueName: \"kubernetes.io/projected/f371d8f0-5386-4a69-b81c-8c656d71cd54-kube-api-access-wgmc4\") pod \"machine-config-controller-84d6567774-54rxk\" (UID: \"f371d8f0-5386-4a69-b81c-8c656d71cd54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" Oct 01 16:01:36 crc kubenswrapper[4726]: W1001 16:01:36.872273 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f008910_43f3_4260_86e6_4cd5a92e28c0.slice/crio-5a48504b0340ba4de61a8702080a186668cb4ff6a08dc3ae2c54ab54a2e4847b WatchSource:0}: Error finding container 5a48504b0340ba4de61a8702080a186668cb4ff6a08dc3ae2c54ab54a2e4847b: Status 404 returned error can't find the container with id 5a48504b0340ba4de61a8702080a186668cb4ff6a08dc3ae2c54ab54a2e4847b Oct 01 16:01:36 crc kubenswrapper[4726]: W1001 16:01:36.875153 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16ec2ea5_2c69_457f_b36f_06011c98666e.slice/crio-8435788b8b1ca75fedd1cc544da1efd868b82bcc9d93e83e7a24f825d55bca2c WatchSource:0}: Error finding container 8435788b8b1ca75fedd1cc544da1efd868b82bcc9d93e83e7a24f825d55bca2c: Status 404 returned error can't find the container with id 8435788b8b1ca75fedd1cc544da1efd868b82bcc9d93e83e7a24f825d55bca2c Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.882411 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mpfpx"] Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.883962 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6k29\" (UniqueName: \"kubernetes.io/projected/e500be09-9f04-49a8-8ad2-e69012bd36f1-kube-api-access-d6k29\") pod \"multus-admission-controller-857f4d67dd-6w44d\" (UID: \"e500be09-9f04-49a8-8ad2-e69012bd36f1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6w44d" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.899515 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:36 crc kubenswrapper[4726]: E1001 16:01:36.900305 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.400283141 +0000 UTC m=+150.301835718 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.902754 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" Oct 01 16:01:36 crc kubenswrapper[4726]: W1001 16:01:36.903300 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5399581d_2111_47d1_bfe7_007b7c46d706.slice/crio-6ef995d358180f4681ba8e75ce05e761699328df0c7d0b4462e5e2f18d96d4c0 WatchSource:0}: Error finding container 6ef995d358180f4681ba8e75ce05e761699328df0c7d0b4462e5e2f18d96d4c0: Status 404 returned error can't find the container with id 6ef995d358180f4681ba8e75ce05e761699328df0c7d0b4462e5e2f18d96d4c0 Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.908412 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgpvw\" (UniqueName: \"kubernetes.io/projected/60142257-8641-4fba-83ca-bb1af9f68c0d-kube-api-access-tgpvw\") pod \"router-default-5444994796-9x846\" (UID: \"60142257-8641-4fba-83ca-bb1af9f68c0d\") " pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.923491 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcp85\" (UniqueName: \"kubernetes.io/projected/8282f47b-cd59-4cba-a1aa-7528d57842ec-kube-api-access-qcp85\") pod \"control-plane-machine-set-operator-78cbb6b69f-z4ml2\" (UID: \"8282f47b-cd59-4cba-a1aa-7528d57842ec\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z4ml2" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.945117 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6zb9\" (UniqueName: \"kubernetes.io/projected/a44f6721-0fe0-45cc-8290-fe52ffb3f8f2-kube-api-access-b6zb9\") pod \"openshift-config-operator-7777fb866f-cg79v\" (UID: \"a44f6721-0fe0-45cc-8290-fe52ffb3f8f2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" Oct 01 16:01:36 crc kubenswrapper[4726]: W1001 16:01:36.956762 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod052caf82_2622_48b3_8c66_02955a93dcf5.slice/crio-44d376b2554487ec795b5aacbf1e88e8d6ec8be2a9a9d9351a42589f1f3d1fc1 WatchSource:0}: Error finding container 44d376b2554487ec795b5aacbf1e88e8d6ec8be2a9a9d9351a42589f1f3d1fc1: Status 404 returned error can't find the container with id 44d376b2554487ec795b5aacbf1e88e8d6ec8be2a9a9d9351a42589f1f3d1fc1 Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.964728 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt7mt\" (UniqueName: \"kubernetes.io/projected/1a8d300c-d03d-4927-b9ba-c17f692a9db4-kube-api-access-nt7mt\") pod \"route-controller-manager-6576b87f9c-2gcb8\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.974104 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" Oct 01 16:01:36 crc kubenswrapper[4726]: I1001 16:01:36.982421 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7tsz\" (UniqueName: \"kubernetes.io/projected/15670655-7e86-4731-bb0b-7ebd96a20e59-kube-api-access-g7tsz\") pod \"openshift-controller-manager-operator-756b6f6bc6-fkxmw\" (UID: \"15670655-7e86-4731-bb0b-7ebd96a20e59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.000508 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:37 crc kubenswrapper[4726]: E1001 16:01:37.000790 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.500758665 +0000 UTC m=+150.402311262 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.001261 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:37 crc kubenswrapper[4726]: E1001 16:01:37.001651 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.501641678 +0000 UTC m=+150.403194265 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.003327 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.006889 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dns5\" (UniqueName: \"kubernetes.io/projected/b8cf6941-f833-4465-a3e9-9efea4b9b5b0-kube-api-access-7dns5\") pod \"downloads-7954f5f757-5tjmt\" (UID: \"b8cf6941-f833-4465-a3e9-9efea4b9b5b0\") " pod="openshift-console/downloads-7954f5f757-5tjmt" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.018642 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.025570 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6w44d" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.033337 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82hfx\" (UniqueName: \"kubernetes.io/projected/a79834ba-7c32-4c45-a927-66cc2ebf3e97-kube-api-access-82hfx\") pod \"cluster-samples-operator-665b6dd947-k6p79\" (UID: \"a79834ba-7c32-4c45-a927-66cc2ebf3e97\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.046817 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x82vr\" (UniqueName: \"kubernetes.io/projected/79d896b5-b34d-4206-92d8-99caeafdffb2-kube-api-access-x82vr\") pod \"console-f9d7485db-w4c9z\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.066601 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmj5x\" (UniqueName: \"kubernetes.io/projected/5450e1a6-a528-4242-8d81-a3241131dcc5-kube-api-access-vmj5x\") pod \"service-ca-operator-777779d784-kjhjx\" (UID: \"5450e1a6-a528-4242-8d81-a3241131dcc5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.075867 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.085833 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqkk7\" (UniqueName: \"kubernetes.io/projected/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-kube-api-access-mqkk7\") pod \"collect-profiles-29322240-5zpj4\" (UID: \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.090952 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.103455 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.104024 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:37 crc kubenswrapper[4726]: E1001 16:01:37.104114 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.604075384 +0000 UTC m=+150.505627991 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.115167 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.116738 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trjk7\" (UniqueName: \"kubernetes.io/projected/8db32339-b3a2-4b00-9352-575da2340e03-kube-api-access-trjk7\") pod \"packageserver-d55dfcdfc-6sm6z\" (UID: \"8db32339-b3a2-4b00-9352-575da2340e03\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.129146 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zl4l\" (UniqueName: \"kubernetes.io/projected/afecdb5b-56b5-4112-9e71-5f05e43302f0-kube-api-access-5zl4l\") pod \"catalog-operator-68c6474976-w6hhh\" (UID: \"afecdb5b-56b5-4112-9e71-5f05e43302f0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.132260 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.151786 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqg5p\" (UniqueName: \"kubernetes.io/projected/4739e7eb-ea30-427f-a701-29eb99ddd136-kube-api-access-nqg5p\") pod \"dns-default-l6d7x\" (UID: \"4739e7eb-ea30-427f-a701-29eb99ddd136\") " pod="openshift-dns/dns-default-l6d7x" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.185970 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.194456 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.194898 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx8qg\" (UniqueName: \"kubernetes.io/projected/c442dae3-d66d-47a2-b1dc-30332e947203-kube-api-access-dx8qg\") pod \"marketplace-operator-79b997595-9qvwb\" (UID: \"c442dae3-d66d-47a2-b1dc-30332e947203\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.204790 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.204874 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhwlw\" (UniqueName: \"kubernetes.io/projected/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-kube-api-access-jhwlw\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:37 crc kubenswrapper[4726]: E1001 16:01:37.205491 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.705473372 +0000 UTC m=+150.607025949 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.211141 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z4ml2" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.211927 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhwlw\" (UniqueName: \"kubernetes.io/projected/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-kube-api-access-jhwlw\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.219456 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-5tjmt" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.222384 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv"] Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.222478 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jgvx\" (UniqueName: \"kubernetes.io/projected/6d58f9c2-068f-4592-8de6-5217d0f51742-kube-api-access-2jgvx\") pod \"kube-storage-version-migrator-operator-b67b599dd-ns7lp\" (UID: \"6d58f9c2-068f-4592-8de6-5217d0f51742\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.223795 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/939458a3-4eaf-46eb-aeb5-be7daf4f0f8c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bnh2z\" (UID: \"939458a3-4eaf-46eb-aeb5-be7daf4f0f8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.227461 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5zsg\" (UniqueName: \"kubernetes.io/projected/027579b0-86e0-4143-a1ad-8db0c669b7a2-kube-api-access-n5zsg\") pod \"migrator-59844c95c7-7rq8t\" (UID: \"027579b0-86e0-4143-a1ad-8db0c669b7a2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rq8t" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.257253 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f9kw\" (UniqueName: \"kubernetes.io/projected/960e1b01-454c-4563-ba02-c35754a0b970-kube-api-access-7f9kw\") pod \"package-server-manager-789f6589d5-szpqh\" (UID: \"960e1b01-454c-4563-ba02-c35754a0b970\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.273481 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98ea4e5f-eaa2-4612-81a5-5064a6dbc957-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-p7r2h\" (UID: \"98ea4e5f-eaa2-4612-81a5-5064a6dbc957\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.288662 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dg6w\" (UniqueName: \"kubernetes.io/projected/7ab8a342-ec8c-4466-bfaa-d86e384736d8-kube-api-access-8dg6w\") pod \"machine-config-server-6p58q\" (UID: \"7ab8a342-ec8c-4466-bfaa-d86e384736d8\") " pod="openshift-machine-config-operator/machine-config-server-6p58q" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.307758 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:37 crc kubenswrapper[4726]: E1001 16:01:37.308122 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.808083442 +0000 UTC m=+150.709636019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.308310 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.313323 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqfmk\" (UniqueName: \"kubernetes.io/projected/5b28caec-1680-4b2f-a530-644790ea6d7e-kube-api-access-kqfmk\") pod \"olm-operator-6b444d44fb-f7tfz\" (UID: \"5b28caec-1680-4b2f-a530-644790ea6d7e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" Oct 01 16:01:37 crc kubenswrapper[4726]: E1001 16:01:37.326411 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.80872894 +0000 UTC m=+150.710281517 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.332304 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.340570 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts2ck\" (UniqueName: \"kubernetes.io/projected/635021dd-a4fb-4bb2-ba4b-bbce8b29856f-kube-api-access-ts2ck\") pod \"csi-hostpathplugin-j96mp\" (UID: \"635021dd-a4fb-4bb2-ba4b-bbce8b29856f\") " pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.340755 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.351510 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.356707 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tqlj\" (UniqueName: \"kubernetes.io/projected/71c3e368-7f4a-43e1-94a6-81374ba2fa00-kube-api-access-8tqlj\") pod \"service-ca-9c57cc56f-4r5zz\" (UID: \"71c3e368-7f4a-43e1-94a6-81374ba2fa00\") " pod="openshift-service-ca/service-ca-9c57cc56f-4r5zz" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.356950 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.364940 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-4r5zz" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.366793 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6w44d"] Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.369253 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.377330 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.378461 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7n5q\" (UniqueName: \"kubernetes.io/projected/c421f8b0-bb08-4081-8071-9a5f7546630f-kube-api-access-l7n5q\") pod \"etcd-operator-b45778765-b842d\" (UID: \"c421f8b0-bb08-4081-8071-9a5f7546630f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.384860 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.385810 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4bds\" (UniqueName: \"kubernetes.io/projected/e7f82b60-0afe-442d-ab72-40411e685fbc-kube-api-access-m4bds\") pod \"ingress-canary-5jsl4\" (UID: \"e7f82b60-0afe-442d-ab72-40411e685fbc\") " pod="openshift-ingress-canary/ingress-canary-5jsl4" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.399271 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.404353 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.409602 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:37 crc kubenswrapper[4726]: E1001 16:01:37.409877 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.90984191 +0000 UTC m=+150.811394527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.410104 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:37 crc kubenswrapper[4726]: E1001 16:01:37.410471 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:37.910462397 +0000 UTC m=+150.812014974 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.415319 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.418842 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rq8t" Oct 01 16:01:37 crc kubenswrapper[4726]: W1001 16:01:37.418962 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode500be09_9f04_49a8_8ad2_e69012bd36f1.slice/crio-269a493e863e0d6b0d01421037633da2e255f56c33196e608ebde55124a6930e WatchSource:0}: Error finding container 269a493e863e0d6b0d01421037633da2e255f56c33196e608ebde55124a6930e: Status 404 returned error can't find the container with id 269a493e863e0d6b0d01421037633da2e255f56c33196e608ebde55124a6930e Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.425357 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk"] Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.425855 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-5jsl4" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.434191 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-l6d7x" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.450936 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-j96mp" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.457499 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6p58q" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.512044 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:37 crc kubenswrapper[4726]: E1001 16:01:37.513150 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:38.013033676 +0000 UTC m=+150.914586243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.570868 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" event={"ID":"875b60c0-2704-4bb1-9b01-5e6b0c745fa1","Type":"ContainerStarted","Data":"79a70b9e2928f8d514e3032051c54621ab0bf490f16f6c0610ff45a767b6b90d"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.570940 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" event={"ID":"875b60c0-2704-4bb1-9b01-5e6b0c745fa1","Type":"ContainerStarted","Data":"b5937e733d4e1b1522e2779ff812f6aa8291624141c029958191cacab34d72cc"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.571124 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.575204 4726 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-pq5pm container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" start-of-body= Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.575252 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" podUID="875b60c0-2704-4bb1-9b01-5e6b0c745fa1" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.575319 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr" event={"ID":"7f189676-cdd8-4138-b73f-863e1c075bc2","Type":"ContainerStarted","Data":"36ac0f4550ac76dc959aa414f6932db908ab3d7ba81095846b92f310997d939b"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.576357 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9x846" event={"ID":"60142257-8641-4fba-83ca-bb1af9f68c0d","Type":"ContainerStarted","Data":"3f484c2667bab3d3907596bdc41373d3ae46931a27856e1f2b2df7214e11eaeb"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.577578 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" event={"ID":"10a86730-3fae-4053-b4e0-48effedbf3de","Type":"ContainerStarted","Data":"c8bda219734864faf75400b8e227d11a2ea882820696ed507f667fbd7e7a6590"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.579570 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-np6df" event={"ID":"51675697-3e42-4381-a278-44377e79479e","Type":"ContainerStarted","Data":"9e74d74f206a273900c9bba091168191a7851e72cabc0de439c9f32f106b24d3"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.579607 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-np6df" event={"ID":"51675697-3e42-4381-a278-44377e79479e","Type":"ContainerStarted","Data":"6aecd2b0323611143d017ec32cd0cd0301535972a5e9885477b792c3d62666c4"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.582691 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-7jshc" event={"ID":"654d70bb-a9aa-48d1-99dd-95a50b221440","Type":"ContainerStarted","Data":"ef6dc4338dd34e0dbfca9f9a8466ee0830d161275602b2b4277af1955a0330df"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.582735 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-7jshc" event={"ID":"654d70bb-a9aa-48d1-99dd-95a50b221440","Type":"ContainerStarted","Data":"0563a05056f5e7c7913ea4a1f0124476d0df838114ec618fc985717162ccabed"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.582887 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.585886 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" event={"ID":"25ce2f03-84c1-496b-9dbc-c8efba41e3d3","Type":"ContainerStarted","Data":"1c0c1ff467aa81d048d2ffc9f4090a9631ea2e61c7b93cd79c8e4e52affccbe7"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.585917 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" event={"ID":"25ce2f03-84c1-496b-9dbc-c8efba41e3d3","Type":"ContainerStarted","Data":"d33fdd02b7c4ae1469e9b7dbccc2d00cc984ad2d683d00df27ac7892dacf3d74"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.586773 4726 patch_prober.go:28] interesting pod/console-operator-58897d9998-7jshc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.586818 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-7jshc" podUID="654d70bb-a9aa-48d1-99dd-95a50b221440" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.587293 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558" event={"ID":"2f008910-43f3-4260-86e6-4cd5a92e28c0","Type":"ContainerStarted","Data":"5a48504b0340ba4de61a8702080a186668cb4ff6a08dc3ae2c54ab54a2e4847b"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.589236 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" event={"ID":"73789e0a-a83d-4470-9f4c-137b3c317b85","Type":"ContainerStarted","Data":"1130beea1319e8fa0fc79c765ffa78d587a366351938b179f4cc04bd68f8238e"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.589284 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" event={"ID":"73789e0a-a83d-4470-9f4c-137b3c317b85","Type":"ContainerStarted","Data":"28894c50be0b5ef619576dd433216ee43b3d48b8cf5f6b9065b21a32fb43d9fb"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.589446 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.590803 4726 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-zhfrh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.590927 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6w44d" event={"ID":"e500be09-9f04-49a8-8ad2-e69012bd36f1","Type":"ContainerStarted","Data":"269a493e863e0d6b0d01421037633da2e255f56c33196e608ebde55124a6930e"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.590927 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" podUID="73789e0a-a83d-4470-9f4c-137b3c317b85" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.593263 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" event={"ID":"16ec2ea5-2c69-457f-b36f-06011c98666e","Type":"ContainerStarted","Data":"7a65b49ab530183f98cd75c05b2b21e52bfdd342c50df201a89e3ac7b58b15bd"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.593301 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" event={"ID":"16ec2ea5-2c69-457f-b36f-06011c98666e","Type":"ContainerStarted","Data":"69c023f2c136d0eb2459dda1c342de01aaab9bfa0409547c2b8f07cf0b4c1ad6"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.593341 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" event={"ID":"16ec2ea5-2c69-457f-b36f-06011c98666e","Type":"ContainerStarted","Data":"8435788b8b1ca75fedd1cc544da1efd868b82bcc9d93e83e7a24f825d55bca2c"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.595731 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" event={"ID":"5399581d-2111-47d1-bfe7-007b7c46d706","Type":"ContainerStarted","Data":"2fc1094a386cad035a455cdaaac4ae3c8e2129cc47b07e4d31dfc82038fa5e1e"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.595764 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" event={"ID":"5399581d-2111-47d1-bfe7-007b7c46d706","Type":"ContainerStarted","Data":"6ef995d358180f4681ba8e75ce05e761699328df0c7d0b4462e5e2f18d96d4c0"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.600954 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" event={"ID":"052caf82-2622-48b3-8c66-02955a93dcf5","Type":"ContainerStarted","Data":"44d376b2554487ec795b5aacbf1e88e8d6ec8be2a9a9d9351a42589f1f3d1fc1"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.604881 4726 generic.go:334] "Generic (PLEG): container finished" podID="262d42bb-7bb1-49f1-8cf0-ad225e6dd673" containerID="9eaf9f245c6e3e0ce5cff908753a04d64f4ea536886cf83c75c170bcbf232894" exitCode=0 Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.604949 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" event={"ID":"262d42bb-7bb1-49f1-8cf0-ad225e6dd673","Type":"ContainerDied","Data":"9eaf9f245c6e3e0ce5cff908753a04d64f4ea536886cf83c75c170bcbf232894"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.607776 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" event={"ID":"f371d8f0-5386-4a69-b81c-8c656d71cd54","Type":"ContainerStarted","Data":"8d184d61bfe566de8c16551b9820268a7d40edc8531fce5049833d0008e936ba"} Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.613872 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:37 crc kubenswrapper[4726]: E1001 16:01:37.614152 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:38.114140136 +0000 UTC m=+151.015692713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.690806 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pxlbv"] Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.691974 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx"] Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.716877 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:37 crc kubenswrapper[4726]: E1001 16:01:37.717250 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:38.21723278 +0000 UTC m=+151.118785367 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.721289 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.721331 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-audit\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.721425 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-config\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.723972 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-config\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.724757 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f-audit\") pod \"apiserver-76f77b778f-vxgvr\" (UID: \"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f\") " pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.729150 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:37 crc kubenswrapper[4726]: E1001 16:01:37.729843 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:38.229811884 +0000 UTC m=+151.131364461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.773771 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.835087 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.835373 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.835410 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.835437 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:37 crc kubenswrapper[4726]: E1001 16:01:37.836488 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:38.336470562 +0000 UTC m=+151.238023139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.846347 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.849827 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.851412 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.941943 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:37 crc kubenswrapper[4726]: E1001 16:01:37.942552 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:38.442538975 +0000 UTC m=+151.344091552 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:37 crc kubenswrapper[4726]: I1001 16:01:37.965305 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.029758 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.045294 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:38 crc kubenswrapper[4726]: E1001 16:01:38.045428 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:38.545402682 +0000 UTC m=+151.446955259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.046287 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:38 crc kubenswrapper[4726]: E1001 16:01:38.046980 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:38.546962693 +0000 UTC m=+151.448515260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.126413 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.136594 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.149949 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:38 crc kubenswrapper[4726]: E1001 16:01:38.150279 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:38.650260022 +0000 UTC m=+151.551812599 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.250901 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:38 crc kubenswrapper[4726]: E1001 16:01:38.251995 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:38.751984059 +0000 UTC m=+151.653536636 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.344382 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-7jshc" podStartSLOduration=128.344363537 podStartE2EDuration="2m8.344363537s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:38.343184255 +0000 UTC m=+151.244736842" watchObservedRunningTime="2025-10-01 16:01:38.344363537 +0000 UTC m=+151.245916114" Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.352082 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:38 crc kubenswrapper[4726]: E1001 16:01:38.355927 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:38.855903094 +0000 UTC m=+151.757455671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.356026 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:38 crc kubenswrapper[4726]: E1001 16:01:38.356427 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:38.856417648 +0000 UTC m=+151.757970225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.447324 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2sgp2" podStartSLOduration=128.447304996 podStartE2EDuration="2m8.447304996s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:38.446505355 +0000 UTC m=+151.348057942" watchObservedRunningTime="2025-10-01 16:01:38.447304996 +0000 UTC m=+151.348857573" Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.459827 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:38 crc kubenswrapper[4726]: E1001 16:01:38.460106 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:38.960091266 +0000 UTC m=+151.861643843 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.560507 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:38 crc kubenswrapper[4726]: E1001 16:01:38.560859 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:39.060846127 +0000 UTC m=+151.962398694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.627548 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9x846" event={"ID":"60142257-8641-4fba-83ca-bb1af9f68c0d","Type":"ContainerStarted","Data":"cdc3b3429772bcb0cc4ae68ca6fe5649b16169eddb2140f0b05ca6664c6645eb"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.643636 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558" event={"ID":"2f008910-43f3-4260-86e6-4cd5a92e28c0","Type":"ContainerStarted","Data":"ec597122f5aa041c5fe74f2c6ba4c04e43475d4eb136d39b910ed1ed0d16fee3"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.650656 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" podStartSLOduration=128.650637116 podStartE2EDuration="2m8.650637116s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:38.650503913 +0000 UTC m=+151.552056490" watchObservedRunningTime="2025-10-01 16:01:38.650637116 +0000 UTC m=+151.552189693" Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.658618 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6p58q" event={"ID":"7ab8a342-ec8c-4466-bfaa-d86e384736d8","Type":"ContainerStarted","Data":"9542bb11414b2a9b86b7d21f1f8729f5465a40a126ef5587e976f2749354374f"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.658664 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6p58q" event={"ID":"7ab8a342-ec8c-4466-bfaa-d86e384736d8","Type":"ContainerStarted","Data":"f0d22530126f8d3d3256c8cadb839bc04db6d8aaaa5cc0500b77d4a36db3f039"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.660835 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw"] Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.661099 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:38 crc kubenswrapper[4726]: E1001 16:01:38.661331 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:39.16131107 +0000 UTC m=+152.062863647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.661755 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:38 crc kubenswrapper[4726]: E1001 16:01:38.663471 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:39.163455727 +0000 UTC m=+152.065008304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.666506 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx" event={"ID":"5450e1a6-a528-4242-8d81-a3241131dcc5","Type":"ContainerStarted","Data":"c2da66c13a8bf65f426982b9c7fa8f7dfe0025506a8cb730f9eef92dce2002f2"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.666544 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx" event={"ID":"5450e1a6-a528-4242-8d81-a3241131dcc5","Type":"ContainerStarted","Data":"45aba9952cf4853b36c5390a43c4e4ea628b0bd0a96172d6726166a2b767c2f7"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.671856 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-w4c9z"] Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.685571 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-np6df" event={"ID":"51675697-3e42-4381-a278-44377e79479e","Type":"ContainerStarted","Data":"4d6c3bc5706d71b256620dccb04a0cd0b8ab2866aaa73f2f5787d2751788fc2f"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.708128 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" event={"ID":"3f188f7e-b855-4902-b172-db25f2a53746","Type":"ContainerStarted","Data":"e7e2e043f7ae108f8110b888395111963af0c2f5a3e27aa11d1760a915ea8ba6"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.708183 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" event={"ID":"3f188f7e-b855-4902-b172-db25f2a53746","Type":"ContainerStarted","Data":"edb26c6e16403a095a6be68c932b01723065e6a61daea87b00528798a71f02eb"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.717722 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cg79v"] Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.727302 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" event={"ID":"052caf82-2622-48b3-8c66-02955a93dcf5","Type":"ContainerStarted","Data":"8c6a4dbec3397748cffd9663af45fc494da1c82d86eeb7a6ef1bff681cf5e8a9"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.727374 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" event={"ID":"052caf82-2622-48b3-8c66-02955a93dcf5","Type":"ContainerStarted","Data":"37a59e2de4f4f8686d25935afb581463ec4372a6bfbb9d5015a51c9f907e610b"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.742997 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" event={"ID":"5399581d-2111-47d1-bfe7-007b7c46d706","Type":"ContainerStarted","Data":"159e923e10d037c0668ef67e7084d90b24797ee9554c8d577b708990547badc3"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.760612 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" event={"ID":"262d42bb-7bb1-49f1-8cf0-ad225e6dd673","Type":"ContainerStarted","Data":"1de3e88b28bc1026cfbc9e31e50973894f6be3431eb41f53b4f50e5d6adf2dc8"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.762210 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:38 crc kubenswrapper[4726]: E1001 16:01:38.762750 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:39.262728569 +0000 UTC m=+152.164281146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.765224 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8"] Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.772007 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79"] Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.772490 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" event={"ID":"f371d8f0-5386-4a69-b81c-8c656d71cd54","Type":"ContainerStarted","Data":"0710eb440fe783c54f9f63cc030ddb9fd43494b166b23086c5fe5af35c970021"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.772520 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" event={"ID":"f371d8f0-5386-4a69-b81c-8c656d71cd54","Type":"ContainerStarted","Data":"871d052641904ceec5fd78d3b95ab27b3e157eb17696cab6fe3a34b158cfd4c8"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.783640 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6w44d" event={"ID":"e500be09-9f04-49a8-8ad2-e69012bd36f1","Type":"ContainerStarted","Data":"533ded5a809939f204ab7ad384a5873f42ecd5f598779874523b0a2ed0d96a85"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.784322 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" podStartSLOduration=128.784306423 podStartE2EDuration="2m8.784306423s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:38.781975181 +0000 UTC m=+151.683527748" watchObservedRunningTime="2025-10-01 16:01:38.784306423 +0000 UTC m=+151.685859000" Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.808805 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" event={"ID":"10a86730-3fae-4053-b4e0-48effedbf3de","Type":"ContainerStarted","Data":"7ffd88bf1dd563911048a477b50d752a838d04e5de1199efe0082aaacc814981"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.808852 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" event={"ID":"10a86730-3fae-4053-b4e0-48effedbf3de","Type":"ContainerStarted","Data":"0cf1af45d70e24e0fb8b93782eab99c41a4e83c1c4539c154ee6003c4be15501"} Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.828078 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.829738 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rgfgr" podStartSLOduration=128.829723822 podStartE2EDuration="2m8.829723822s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:38.827478932 +0000 UTC m=+151.729031509" watchObservedRunningTime="2025-10-01 16:01:38.829723822 +0000 UTC m=+151.731276389" Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.863306 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:38 crc kubenswrapper[4726]: E1001 16:01:38.865392 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:39.36537545 +0000 UTC m=+152.266928027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.934842 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:01:38 crc kubenswrapper[4726]: I1001 16:01:38.964921 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:38 crc kubenswrapper[4726]: E1001 16:01:38.966073 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:39.466022318 +0000 UTC m=+152.367574935 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.019392 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.025507 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b62jx" podStartSLOduration=129.025491651 podStartE2EDuration="2m9.025491651s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:39.025199563 +0000 UTC m=+151.926752140" watchObservedRunningTime="2025-10-01 16:01:39.025491651 +0000 UTC m=+151.927044228" Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.066799 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:39 crc kubenswrapper[4726]: E1001 16:01:39.067425 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:39.567404966 +0000 UTC m=+152.468957553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.076005 4726 patch_prober.go:28] interesting pod/router-default-5444994796-9x846 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 16:01:39 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 01 16:01:39 crc kubenswrapper[4726]: [+]process-running ok Oct 01 16:01:39 crc kubenswrapper[4726]: healthz check failed Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.076104 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9x846" podUID="60142257-8641-4fba-83ca-bb1af9f68c0d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.090695 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh"] Oct 01 16:01:39 crc kubenswrapper[4726]: W1001 16:01:39.122887 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod960e1b01_454c_4563_ba02_c35754a0b970.slice/crio-ca197b848cc54b08bed13d34418559ac2e7d6eed94ce370d596b08090c800356 WatchSource:0}: Error finding container ca197b848cc54b08bed13d34418559ac2e7d6eed94ce370d596b08090c800356: Status 404 returned error can't find the container with id ca197b848cc54b08bed13d34418559ac2e7d6eed94ce370d596b08090c800356 Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.122985 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.130461 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-5tjmt"] Oct 01 16:01:39 crc kubenswrapper[4726]: W1001 16:01:39.139393 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d58f9c2_068f_4592_8de6_5217d0f51742.slice/crio-9646874056a07b7ee52e9470d46023ec28e307bc936d8731fc157d644bf91fb3 WatchSource:0}: Error finding container 9646874056a07b7ee52e9470d46023ec28e307bc936d8731fc157d644bf91fb3: Status 404 returned error can't find the container with id 9646874056a07b7ee52e9470d46023ec28e307bc936d8731fc157d644bf91fb3 Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.152957 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4r5zz"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.171929 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:39 crc kubenswrapper[4726]: E1001 16:01:39.172241 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:39.672225805 +0000 UTC m=+152.573778392 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.188514 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-7jshc" Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.273296 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:39 crc kubenswrapper[4726]: E1001 16:01:39.273864 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:39.773841279 +0000 UTC m=+152.675393856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.305791 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-6p58q" podStartSLOduration=5.305775809 podStartE2EDuration="5.305775809s" podCreationTimestamp="2025-10-01 16:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:39.304264269 +0000 UTC m=+152.205816856" watchObservedRunningTime="2025-10-01 16:01:39.305775809 +0000 UTC m=+152.207328386" Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.374988 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:39 crc kubenswrapper[4726]: E1001 16:01:39.375307 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:39.875291518 +0000 UTC m=+152.776844095 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.400305 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" podStartSLOduration=129.400274373 podStartE2EDuration="2m9.400274373s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:39.397591432 +0000 UTC m=+152.299144009" watchObservedRunningTime="2025-10-01 16:01:39.400274373 +0000 UTC m=+152.301826950" Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.476306 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:39 crc kubenswrapper[4726]: E1001 16:01:39.476619 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:39.976603454 +0000 UTC m=+152.878156031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.486447 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-6w44d" podStartSLOduration=129.486431756 podStartE2EDuration="2m9.486431756s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:39.482993494 +0000 UTC m=+152.384546071" watchObservedRunningTime="2025-10-01 16:01:39.486431756 +0000 UTC m=+152.387984333" Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.566076 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjhjx" podStartSLOduration=129.566026754 podStartE2EDuration="2m9.566026754s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:39.508548194 +0000 UTC m=+152.410100781" watchObservedRunningTime="2025-10-01 16:01:39.566026754 +0000 UTC m=+152.467579331" Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.571207 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qvwb"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.571490 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-np6df" podStartSLOduration=129.571472279 podStartE2EDuration="2m9.571472279s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:39.554096006 +0000 UTC m=+152.455648593" watchObservedRunningTime="2025-10-01 16:01:39.571472279 +0000 UTC m=+152.473024856" Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.577550 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:39 crc kubenswrapper[4726]: E1001 16:01:39.577879 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:40.077862509 +0000 UTC m=+152.979415086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.596715 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh"] Oct 01 16:01:39 crc kubenswrapper[4726]: W1001 16:01:39.599192 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc442dae3_d66d_47a2_b1dc_30332e947203.slice/crio-0e73fc792d3c86d3340a0208a3f4725c6f8e7a9dd4082e7049870f04bbed0876 WatchSource:0}: Error finding container 0e73fc792d3c86d3340a0208a3f4725c6f8e7a9dd4082e7049870f04bbed0876: Status 404 returned error can't find the container with id 0e73fc792d3c86d3340a0208a3f4725c6f8e7a9dd4082e7049870f04bbed0876 Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.608462 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m4w6w" podStartSLOduration=129.608439322 podStartE2EDuration="2m9.608439322s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:39.593566017 +0000 UTC m=+152.495118594" watchObservedRunningTime="2025-10-01 16:01:39.608439322 +0000 UTC m=+152.509991899" Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.631415 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-67558" podStartSLOduration=129.631394553 podStartE2EDuration="2m9.631394553s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:39.629524683 +0000 UTC m=+152.531077270" watchObservedRunningTime="2025-10-01 16:01:39.631394553 +0000 UTC m=+152.532947140" Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.679937 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:39 crc kubenswrapper[4726]: E1001 16:01:39.680573 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:40.180559971 +0000 UTC m=+153.082112538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.724113 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.724714 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.752142 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.754107 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-9x846" podStartSLOduration=129.754091408 podStartE2EDuration="2m9.754091408s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:39.730449559 +0000 UTC m=+152.632002136" watchObservedRunningTime="2025-10-01 16:01:39.754091408 +0000 UTC m=+152.655643985" Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.755723 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-b842d"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.783662 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:39 crc kubenswrapper[4726]: E1001 16:01:39.784069 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:40.284038585 +0000 UTC m=+153.185591162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:39 crc kubenswrapper[4726]: W1001 16:01:39.785731 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98ea4e5f_eaa2_4612_81a5_5064a6dbc957.slice/crio-607f312356e20dca3e3e1c37c231dd4be290cb5a87674b81a33708f9b244f3a5 WatchSource:0}: Error finding container 607f312356e20dca3e3e1c37c231dd4be290cb5a87674b81a33708f9b244f3a5: Status 404 returned error can't find the container with id 607f312356e20dca3e3e1c37c231dd4be290cb5a87674b81a33708f9b244f3a5 Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.786965 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7rq8t"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.803211 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.803269 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-j96mp"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.817872 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-pxlbv" podStartSLOduration=129.817852984 podStartE2EDuration="2m9.817852984s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:39.803817191 +0000 UTC m=+152.705369768" watchObservedRunningTime="2025-10-01 16:01:39.817852984 +0000 UTC m=+152.719405571" Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.858665 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-54rxk" podStartSLOduration=129.8586504 podStartE2EDuration="2m9.8586504s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:39.856561175 +0000 UTC m=+152.758113752" watchObservedRunningTime="2025-10-01 16:01:39.8586504 +0000 UTC m=+152.760202967" Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.885586 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:39 crc kubenswrapper[4726]: E1001 16:01:39.885939 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:40.385923135 +0000 UTC m=+153.287475712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.890134 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-l6d7x"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.890164 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-4r5zz" event={"ID":"71c3e368-7f4a-43e1-94a6-81374ba2fa00","Type":"ContainerStarted","Data":"477e08112a9b9b8c7147edaf845c649a78ad42fa808d6fc1a6a58b944288fb73"} Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.890185 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.890201 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z4ml2"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.890214 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-4r5zz" event={"ID":"71c3e368-7f4a-43e1-94a6-81374ba2fa00","Type":"ContainerStarted","Data":"85af40e6ad92320530145503640180384c22c6c054e49bdbe6ed80f9fac07bf7"} Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.901610 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79" event={"ID":"a79834ba-7c32-4c45-a927-66cc2ebf3e97","Type":"ContainerStarted","Data":"a75d6f5fad190bd4c78607d52dc28bade41d01312c698ac88f0eae6105dacaef"} Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.901656 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79" event={"ID":"a79834ba-7c32-4c45-a927-66cc2ebf3e97","Type":"ContainerStarted","Data":"938c875e1e581d06dea82c4d2123706b1499903de0b1893ef6b043f0de35a79a"} Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.904413 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" event={"ID":"c442dae3-d66d-47a2-b1dc-30332e947203","Type":"ContainerStarted","Data":"0e73fc792d3c86d3340a0208a3f4725c6f8e7a9dd4082e7049870f04bbed0876"} Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.909731 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" event={"ID":"a44f6721-0fe0-45cc-8290-fe52ffb3f8f2","Type":"ContainerStarted","Data":"04e6007c2615dc35b4567f37d6d151215885d25e13d6aad125bdef47e2335a4f"} Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.909772 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" event={"ID":"a44f6721-0fe0-45cc-8290-fe52ffb3f8f2","Type":"ContainerStarted","Data":"760fb70a38cb5b5e6ef77d78695943129a8ab158d62077c2a2dfd90545f6c34a"} Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.914725 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw" event={"ID":"15670655-7e86-4731-bb0b-7ebd96a20e59","Type":"ContainerStarted","Data":"9328c7c64de43d14263e8dd0fd47e8ddc22ba1522104622517401dcf9d75057a"} Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.914766 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw" event={"ID":"15670655-7e86-4731-bb0b-7ebd96a20e59","Type":"ContainerStarted","Data":"be7121fc3e0ccd92ef9fff6e42abc1aba9a1808d05c4a8ad7e4e88a620d3b3e5"} Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.920706 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-w4c9z" event={"ID":"79d896b5-b34d-4206-92d8-99caeafdffb2","Type":"ContainerStarted","Data":"462d230384168689aa43fc8e1ca8f9835c9bd43481602a2559c9d7c9f58e35ab"} Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.920759 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-w4c9z" event={"ID":"79d896b5-b34d-4206-92d8-99caeafdffb2","Type":"ContainerStarted","Data":"86c2278e12d739cff7dc0245e601e8400ceb9f8d6cd49937c6ef877acd41a21b"} Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.956798 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-5jsl4"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.962087 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-vxgvr"] Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.969910 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6w44d" event={"ID":"e500be09-9f04-49a8-8ad2-e69012bd36f1","Type":"ContainerStarted","Data":"ec8ddac02fe9d26c9454a13495cb5c7aa2fbd886dc8886f99567744401367f97"} Oct 01 16:01:39 crc kubenswrapper[4726]: I1001 16:01:39.991316 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:39 crc kubenswrapper[4726]: E1001 16:01:39.992305 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:40.492287475 +0000 UTC m=+153.393840052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.003445 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xh5hv" podStartSLOduration=130.003429192 podStartE2EDuration="2m10.003429192s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:39.98946761 +0000 UTC m=+152.891020187" watchObservedRunningTime="2025-10-01 16:01:40.003429192 +0000 UTC m=+152.904981769" Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.070323 4726 patch_prober.go:28] interesting pod/router-default-5444994796-9x846 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 16:01:40 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 01 16:01:40 crc kubenswrapper[4726]: [+]process-running ok Oct 01 16:01:40 crc kubenswrapper[4726]: healthz check failed Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.070687 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9x846" podUID="60142257-8641-4fba-83ca-bb1af9f68c0d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.093444 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:40 crc kubenswrapper[4726]: E1001 16:01:40.095406 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:40.595393759 +0000 UTC m=+153.496946336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.105861 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-mpfpx" podStartSLOduration=130.105839337 podStartE2EDuration="2m10.105839337s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:40.022328594 +0000 UTC m=+152.923881171" watchObservedRunningTime="2025-10-01 16:01:40.105839337 +0000 UTC m=+153.007391924" Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.147300 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh" event={"ID":"960e1b01-454c-4563-ba02-c35754a0b970","Type":"ContainerStarted","Data":"c302be5c49ebbf43ad85d386cec649307a76e57155765559d9ab6ce34d3a25fa"} Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.147336 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh" event={"ID":"960e1b01-454c-4563-ba02-c35754a0b970","Type":"ContainerStarted","Data":"ca197b848cc54b08bed13d34418559ac2e7d6eed94ce370d596b08090c800356"} Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.163256 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fkxmw" podStartSLOduration=130.163239274 podStartE2EDuration="2m10.163239274s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:40.115868663 +0000 UTC m=+153.017421240" watchObservedRunningTime="2025-10-01 16:01:40.163239274 +0000 UTC m=+153.064791851" Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.179107 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" event={"ID":"afecdb5b-56b5-4112-9e71-5f05e43302f0","Type":"ContainerStarted","Data":"8bdf33841e681af1d857adad1bbbd9bf3b4fe06a8044ef1c04622368eea435af"} Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.194269 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:40 crc kubenswrapper[4726]: E1001 16:01:40.194571 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:40.694556807 +0000 UTC m=+153.596109384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.201391 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-5tjmt" event={"ID":"b8cf6941-f833-4465-a3e9-9efea4b9b5b0","Type":"ContainerStarted","Data":"01961fceaa9c005e85a927aaebbda96fa4d54fb2066cefcfb9b501207a524ff3"} Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.201428 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-5tjmt" Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.201438 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-5tjmt" event={"ID":"b8cf6941-f833-4465-a3e9-9efea4b9b5b0","Type":"ContainerStarted","Data":"47522e4f825d0631ced8444e02c370382832ff69358767f21487cfaa91165934"} Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.228519 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp" event={"ID":"6d58f9c2-068f-4592-8de6-5217d0f51742","Type":"ContainerStarted","Data":"5e93151a06e53ecad2515cba168971e4243605ef017be9552bd985bd7b456ef3"} Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.228555 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp" event={"ID":"6d58f9c2-068f-4592-8de6-5217d0f51742","Type":"ContainerStarted","Data":"9646874056a07b7ee52e9470d46023ec28e307bc936d8731fc157d644bf91fb3"} Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.264341 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-5tjmt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.264395 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5tjmt" podUID="b8cf6941-f833-4465-a3e9-9efea4b9b5b0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.264958 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" event={"ID":"1a8d300c-d03d-4927-b9ba-c17f692a9db4","Type":"ContainerStarted","Data":"dfb16ccd3f91ddf9f6fd4194242c4a779dc999ad3abc3556c4666a4c4ed2139d"} Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.264994 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" event={"ID":"1a8d300c-d03d-4927-b9ba-c17f692a9db4","Type":"ContainerStarted","Data":"f6fcd61b312ae9185504efa02bc5a3fe03cbc70d476de06e6702a6037821bcb5"} Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.265010 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.268990 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-w4c9z" podStartSLOduration=130.268977157 podStartE2EDuration="2m10.268977157s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:40.268142945 +0000 UTC m=+153.169695522" watchObservedRunningTime="2025-10-01 16:01:40.268977157 +0000 UTC m=+153.170529734" Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.269775 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-4r5zz" podStartSLOduration=130.269771469 podStartE2EDuration="2m10.269771469s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:40.23035035 +0000 UTC m=+153.131902927" watchObservedRunningTime="2025-10-01 16:01:40.269771469 +0000 UTC m=+153.171324046" Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.299608 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-5tjmt" podStartSLOduration=130.299593112 podStartE2EDuration="2m10.299593112s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:40.297539347 +0000 UTC m=+153.199091924" watchObservedRunningTime="2025-10-01 16:01:40.299593112 +0000 UTC m=+153.201145689" Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.300562 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:40 crc kubenswrapper[4726]: E1001 16:01:40.301315 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:40.801299117 +0000 UTC m=+153.702851694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.302063 4726 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-2gcb8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.302097 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" podUID="1a8d300c-d03d-4927-b9ba-c17f692a9db4" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.377953 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" podStartSLOduration=130.377936787 podStartE2EDuration="2m10.377936787s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:40.339331479 +0000 UTC m=+153.240884056" watchObservedRunningTime="2025-10-01 16:01:40.377936787 +0000 UTC m=+153.279489354" Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.401485 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:40 crc kubenswrapper[4726]: E1001 16:01:40.403207 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:40.903176238 +0000 UTC m=+153.804728815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.505704 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:40 crc kubenswrapper[4726]: E1001 16:01:40.506086 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:41.006075306 +0000 UTC m=+153.907627883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:40 crc kubenswrapper[4726]: E1001 16:01:40.608626 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:41.108602845 +0000 UTC m=+154.010155422 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.609069 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.609367 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:40 crc kubenswrapper[4726]: E1001 16:01:40.609644 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:41.109627852 +0000 UTC m=+154.011180419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.713457 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:40 crc kubenswrapper[4726]: E1001 16:01:40.714017 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:41.213998779 +0000 UTC m=+154.115551356 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.817685 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:40 crc kubenswrapper[4726]: E1001 16:01:40.818193 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:41.318182581 +0000 UTC m=+154.219735158 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:40 crc kubenswrapper[4726]: I1001 16:01:40.919559 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:40 crc kubenswrapper[4726]: E1001 16:01:40.919934 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:41.419914098 +0000 UTC m=+154.321466675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.023583 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:41 crc kubenswrapper[4726]: E1001 16:01:41.023993 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:41.523976907 +0000 UTC m=+154.425529474 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.025098 4726 patch_prober.go:28] interesting pod/router-default-5444994796-9x846 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 16:01:41 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 01 16:01:41 crc kubenswrapper[4726]: [+]process-running ok Oct 01 16:01:41 crc kubenswrapper[4726]: healthz check failed Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.025160 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9x846" podUID="60142257-8641-4fba-83ca-bb1af9f68c0d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.088430 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.090133 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.099253 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.125268 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:41 crc kubenswrapper[4726]: E1001 16:01:41.125690 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:41.625671153 +0000 UTC m=+154.527223730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.136846 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ns7lp" podStartSLOduration=131.13682591 podStartE2EDuration="2m11.13682591s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:40.378214004 +0000 UTC m=+153.279766581" watchObservedRunningTime="2025-10-01 16:01:41.13682591 +0000 UTC m=+154.038378487" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.227017 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:41 crc kubenswrapper[4726]: E1001 16:01:41.227732 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:41.727715348 +0000 UTC m=+154.629267925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.324209 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"8530371d0303e4d02a0565874f83206e8fecbd33b8288b0c6dfa3ef73fc7a50d"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.324252 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"12fdc181542f9bcceeb90d83dbe04f167be3258eef6f3d8fd50d60123e4b4757"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.356073 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:41 crc kubenswrapper[4726]: E1001 16:01:41.356153 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:41.856115855 +0000 UTC m=+154.757668432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.357124 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:41 crc kubenswrapper[4726]: E1001 16:01:41.357606 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:41.857595344 +0000 UTC m=+154.759147921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.368404 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z4ml2" event={"ID":"8282f47b-cd59-4cba-a1aa-7528d57842ec","Type":"ContainerStarted","Data":"a72c3789a67c20ebaa06f94495c592283f5fd2848490f8fc8b014810a1474c68"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.368459 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z4ml2" event={"ID":"8282f47b-cd59-4cba-a1aa-7528d57842ec","Type":"ContainerStarted","Data":"b1582fb3a610d2f28dff94a891c1ea97b697bb236ed76b6f0634bb50392b4cfc"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.416035 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" event={"ID":"f8cbbf14-6dac-425e-9cb6-8579f7a150ee","Type":"ContainerStarted","Data":"821ee00b78fb3620fef1e00d43d963d218a9f1d739df5183f5fd83421f603b7e"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.416298 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" event={"ID":"f8cbbf14-6dac-425e-9cb6-8579f7a150ee","Type":"ContainerStarted","Data":"b08a748bc61d06578024920db7aff0e4aab43fd87461db4491b0b1c0b9146eca"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.434293 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j96mp" event={"ID":"635021dd-a4fb-4bb2-ba4b-bbce8b29856f","Type":"ContainerStarted","Data":"9aeabb378004e73bd91089225102a9fca9f4649a99369f193f33b6e1c87fd2d7"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.459622 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:41 crc kubenswrapper[4726]: E1001 16:01:41.459852 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:41.959831885 +0000 UTC m=+154.861384472 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.459917 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.461175 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh" event={"ID":"960e1b01-454c-4563-ba02-c35754a0b970","Type":"ContainerStarted","Data":"40ac74a6639280e4dcb5f53a028beda259b37842088089b745d7756c48a9017e"} Oct 01 16:01:41 crc kubenswrapper[4726]: E1001 16:01:41.461358 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:41.961334915 +0000 UTC m=+154.862887502 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.461678 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.469413 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z4ml2" podStartSLOduration=131.469397919 podStartE2EDuration="2m11.469397919s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:41.429350414 +0000 UTC m=+154.330902991" watchObservedRunningTime="2025-10-01 16:01:41.469397919 +0000 UTC m=+154.370950496" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.470106 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" podStartSLOduration=101.470100968 podStartE2EDuration="1m41.470100968s" podCreationTimestamp="2025-10-01 16:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:41.469229365 +0000 UTC m=+154.370781942" watchObservedRunningTime="2025-10-01 16:01:41.470100968 +0000 UTC m=+154.371653535" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.489232 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" event={"ID":"c442dae3-d66d-47a2-b1dc-30332e947203","Type":"ContainerStarted","Data":"ac73ff1ef41d1fc551039f1f7f0e4b2e6e4585b60fcfc24370afcf7bd23c6eee"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.489894 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.504258 4726 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9qvwb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.504328 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" podUID="c442dae3-d66d-47a2-b1dc-30332e947203" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.507849 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh" podStartSLOduration=131.507801901 podStartE2EDuration="2m11.507801901s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:41.506802185 +0000 UTC m=+154.408354772" watchObservedRunningTime="2025-10-01 16:01:41.507801901 +0000 UTC m=+154.409354488" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.523776 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"025d4e41ab11d999b80f36fc7edd01c9a821dbd28d906942454b0a0b7218da3b"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.524090 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"8c91d5a4cca87be6349a5499a264e92366d9d96f1e92606a10ff3316ecf1f17d"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.549338 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h" event={"ID":"98ea4e5f-eaa2-4612-81a5-5064a6dbc957","Type":"ContainerStarted","Data":"a567f5ac24e6862228bb94e371b8254451d0192046088266a7a4314798d7a3a2"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.549386 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h" event={"ID":"98ea4e5f-eaa2-4612-81a5-5064a6dbc957","Type":"ContainerStarted","Data":"607f312356e20dca3e3e1c37c231dd4be290cb5a87674b81a33708f9b244f3a5"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.563571 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:41 crc kubenswrapper[4726]: E1001 16:01:41.564891 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:42.06487228 +0000 UTC m=+154.966424857 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.577923 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rq8t" event={"ID":"027579b0-86e0-4143-a1ad-8db0c669b7a2","Type":"ContainerStarted","Data":"537949ee51a4e29e282b87ed3b6137eaba2412685e48a8d252eb4e10a009c6b4"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.577973 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rq8t" event={"ID":"027579b0-86e0-4143-a1ad-8db0c669b7a2","Type":"ContainerStarted","Data":"2f736558d886c20d3168fc26e0c1e522d2b0680f8eaa83e9375c5be47f073eb0"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.577984 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rq8t" event={"ID":"027579b0-86e0-4143-a1ad-8db0c669b7a2","Type":"ContainerStarted","Data":"4f0abafef7dc7e20b790988d03182a7fa8bcc6d447ad9e057f69a479d7776d72"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.603394 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z" event={"ID":"939458a3-4eaf-46eb-aeb5-be7daf4f0f8c","Type":"ContainerStarted","Data":"e3e4480bde50cb35116ad80fedfa99222a54037b4fd5b4365c68a07528c26bda"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.603453 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z" event={"ID":"939458a3-4eaf-46eb-aeb5-be7daf4f0f8c","Type":"ContainerStarted","Data":"2d8bcac2ddcb84acf6c275a695a6cded323d76d579c9b55e68f2fd35b986e594"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.621997 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" podStartSLOduration=131.621977929 podStartE2EDuration="2m11.621977929s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:41.584511372 +0000 UTC m=+154.486063949" watchObservedRunningTime="2025-10-01 16:01:41.621977929 +0000 UTC m=+154.523530506" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.630479 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c987fcf5a4d9ac3d20c404f452cbdd914ea97b1542c98bf534dddfa1c1351cef"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.630516 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d2fa56234b2b2669bcce7b695c8233b8ae0220de74bfece8e6e2b1f71080eabe"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.631132 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.644609 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-p7r2h" podStartSLOduration=131.644593301 podStartE2EDuration="2m11.644593301s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:41.644382615 +0000 UTC m=+154.545935202" watchObservedRunningTime="2025-10-01 16:01:41.644593301 +0000 UTC m=+154.546145868" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.645102 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7rq8t" podStartSLOduration=131.645096375 podStartE2EDuration="2m11.645096375s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:41.623223232 +0000 UTC m=+154.524775829" watchObservedRunningTime="2025-10-01 16:01:41.645096375 +0000 UTC m=+154.546648952" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.649343 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" event={"ID":"8db32339-b3a2-4b00-9352-575da2340e03","Type":"ContainerStarted","Data":"060c5b9b0f215361b1dd23932c4eff7d931db3b9a8ad3d025ac8c01247cfba8d"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.649393 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" event={"ID":"8db32339-b3a2-4b00-9352-575da2340e03","Type":"ContainerStarted","Data":"7914adf4beb38add95f01b1d625bda892341d4dcb0f8eac72a667242546d6e2f"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.650400 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.660120 4726 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-6sm6z container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.660181 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" podUID="8db32339-b3a2-4b00-9352-575da2340e03" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.667066 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:41 crc kubenswrapper[4726]: E1001 16:01:41.667651 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:42.167638154 +0000 UTC m=+155.069190731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.674543 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" event={"ID":"afecdb5b-56b5-4112-9e71-5f05e43302f0","Type":"ContainerStarted","Data":"231c64f32390c9e53b072de779ca2eb59abcbdad7111ac77798b2fd38a8c37c9"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.674589 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.677327 4726 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w6hhh container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.677383 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" podUID="afecdb5b-56b5-4112-9e71-5f05e43302f0" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.698523 4726 generic.go:334] "Generic (PLEG): container finished" podID="a44f6721-0fe0-45cc-8290-fe52ffb3f8f2" containerID="04e6007c2615dc35b4567f37d6d151215885d25e13d6aad125bdef47e2335a4f" exitCode=0 Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.698615 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" event={"ID":"a44f6721-0fe0-45cc-8290-fe52ffb3f8f2","Type":"ContainerDied","Data":"04e6007c2615dc35b4567f37d6d151215885d25e13d6aad125bdef47e2335a4f"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.698644 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" event={"ID":"a44f6721-0fe0-45cc-8290-fe52ffb3f8f2","Type":"ContainerStarted","Data":"884e695924de80a1c399eabc87980b046644cefb67ff6bdda909fd2d2a8cade9"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.699271 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.727405 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" event={"ID":"c421f8b0-bb08-4081-8071-9a5f7546630f","Type":"ContainerStarted","Data":"6d75b8f516330bc58df81cd85a50e55461d20ac6a80cc25411307a30f79b2bfe"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.727471 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" event={"ID":"c421f8b0-bb08-4081-8071-9a5f7546630f","Type":"ContainerStarted","Data":"94de5e58cfd39b80bc9a83f960016e04891fcaa450dd286d2d9695b68e0591f3"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.737184 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" event={"ID":"5b28caec-1680-4b2f-a530-644790ea6d7e","Type":"ContainerStarted","Data":"f6b4183be2d6d442843b57e8d38e1fcb97e4259908df8d0db9be96cd6a181434"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.737242 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" event={"ID":"5b28caec-1680-4b2f-a530-644790ea6d7e","Type":"ContainerStarted","Data":"69497ecb0c6f9c92efc68ef1d5f32590240a0a28b0e787ffb1e53398430d4b4a"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.738217 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.750332 4726 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-f7tfz container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.750679 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" podUID="5b28caec-1680-4b2f-a530-644790ea6d7e" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.752337 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79" event={"ID":"a79834ba-7c32-4c45-a927-66cc2ebf3e97","Type":"ContainerStarted","Data":"6a4782253232f164a00d3472e35bf09e54076b0817cc10fb7ce1d994fabd88c8"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.761952 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" podStartSLOduration=131.761934993 podStartE2EDuration="2m11.761934993s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:41.73287073 +0000 UTC m=+154.634423317" watchObservedRunningTime="2025-10-01 16:01:41.761934993 +0000 UTC m=+154.663487570" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.767684 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:41 crc kubenswrapper[4726]: E1001 16:01:41.768310 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:42.268287322 +0000 UTC m=+155.169839909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.774023 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-l6d7x" event={"ID":"4739e7eb-ea30-427f-a701-29eb99ddd136","Type":"ContainerStarted","Data":"bdd8af200dacef4148f9295791daf2e5fe4bcc8a019d87e7643d90a4c8ce03b9"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.774142 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-l6d7x" event={"ID":"4739e7eb-ea30-427f-a701-29eb99ddd136","Type":"ContainerStarted","Data":"06eee43f500ff1e960c4fa26efcae4067dde3805b48b78c904d811fa5d215dd3"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.774873 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-l6d7x" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.783902 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" event={"ID":"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f","Type":"ContainerStarted","Data":"2a69b46eb9627f2fdd3e07e1fad741887fef5f69d93aaf6c7068b17189251583"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.783942 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" event={"ID":"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f","Type":"ContainerStarted","Data":"380f3f8723fc388b6c14f4026d135c30163130b8afa67d0d338b4e9b4a1e4c4c"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.803294 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bnh2z" podStartSLOduration=131.803276213 podStartE2EDuration="2m11.803276213s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:41.801411604 +0000 UTC m=+154.702964191" watchObservedRunningTime="2025-10-01 16:01:41.803276213 +0000 UTC m=+154.704828790" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.828957 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-5tjmt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.829026 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5tjmt" podUID="b8cf6941-f833-4465-a3e9-9efea4b9b5b0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.846432 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-5jsl4" event={"ID":"e7f82b60-0afe-442d-ab72-40411e685fbc","Type":"ContainerStarted","Data":"6ba4baa58acc3ca8d18991cf1e34b7dc08ff1f4bbbf92f714fe4a590736da3a5"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.846494 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-5jsl4" event={"ID":"e7f82b60-0afe-442d-ab72-40411e685fbc","Type":"ContainerStarted","Data":"99901ad24f7d2731f61f4617ca2183db19bbaa31dcdaa1b168360bcb7aafc128"} Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.855094 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8xtnl" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.856452 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.874879 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.875640 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" podStartSLOduration=131.875621858 podStartE2EDuration="2m11.875621858s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:41.848279481 +0000 UTC m=+154.749832058" watchObservedRunningTime="2025-10-01 16:01:41.875621858 +0000 UTC m=+154.777174435" Oct 01 16:01:41 crc kubenswrapper[4726]: E1001 16:01:41.878888 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:42.378870605 +0000 UTC m=+155.280423182 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.912210 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" podStartSLOduration=131.912191052 podStartE2EDuration="2m11.912191052s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:41.876844981 +0000 UTC m=+154.778397568" watchObservedRunningTime="2025-10-01 16:01:41.912191052 +0000 UTC m=+154.813743629" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.914167 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-l6d7x" podStartSLOduration=7.914154744 podStartE2EDuration="7.914154744s" podCreationTimestamp="2025-10-01 16:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:41.911517854 +0000 UTC m=+154.813070441" watchObservedRunningTime="2025-10-01 16:01:41.914154744 +0000 UTC m=+154.815707321" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.976402 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-b842d" podStartSLOduration=131.97638072 podStartE2EDuration="2m11.97638072s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:41.972874836 +0000 UTC m=+154.874427433" watchObservedRunningTime="2025-10-01 16:01:41.97638072 +0000 UTC m=+154.877933297" Oct 01 16:01:41 crc kubenswrapper[4726]: I1001 16:01:41.978734 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:41 crc kubenswrapper[4726]: E1001 16:01:41.980485 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:42.480456538 +0000 UTC m=+155.382009115 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.023366 4726 patch_prober.go:28] interesting pod/router-default-5444994796-9x846 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 16:01:42 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 01 16:01:42 crc kubenswrapper[4726]: [+]process-running ok Oct 01 16:01:42 crc kubenswrapper[4726]: healthz check failed Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.023633 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9x846" podUID="60142257-8641-4fba-83ca-bb1af9f68c0d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.085430 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:42 crc kubenswrapper[4726]: E1001 16:01:42.085841 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:42.585822602 +0000 UTC m=+155.487375179 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.099872 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k6p79" podStartSLOduration=132.099830214 podStartE2EDuration="2m12.099830214s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:42.039421137 +0000 UTC m=+154.940973714" watchObservedRunningTime="2025-10-01 16:01:42.099830214 +0000 UTC m=+155.001382791" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.101065 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" podStartSLOduration=132.101036686 podStartE2EDuration="2m12.101036686s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:42.099630419 +0000 UTC m=+155.001182996" watchObservedRunningTime="2025-10-01 16:01:42.101036686 +0000 UTC m=+155.002589273" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.136237 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-5jsl4" podStartSLOduration=9.136007417 podStartE2EDuration="9.136007417s" podCreationTimestamp="2025-10-01 16:01:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:42.134160188 +0000 UTC m=+155.035712765" watchObservedRunningTime="2025-10-01 16:01:42.136007417 +0000 UTC m=+155.037559994" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.187149 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:42 crc kubenswrapper[4726]: E1001 16:01:42.187412 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:42.687397334 +0000 UTC m=+155.588949911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.293922 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:42 crc kubenswrapper[4726]: E1001 16:01:42.294402 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:42.794387161 +0000 UTC m=+155.695939738 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.394815 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:42 crc kubenswrapper[4726]: E1001 16:01:42.394952 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:42.894926416 +0000 UTC m=+155.796478993 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.395132 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:42 crc kubenswrapper[4726]: E1001 16:01:42.395409 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:42.895398459 +0000 UTC m=+155.796951036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.496511 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:42 crc kubenswrapper[4726]: E1001 16:01:42.496714 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:42.996675474 +0000 UTC m=+155.898228051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.496828 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:42 crc kubenswrapper[4726]: E1001 16:01:42.497124 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:42.997112446 +0000 UTC m=+155.898665023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.598306 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:42 crc kubenswrapper[4726]: E1001 16:01:42.598530 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:43.098501143 +0000 UTC m=+156.000053720 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.598877 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:42 crc kubenswrapper[4726]: E1001 16:01:42.599208 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:43.099198712 +0000 UTC m=+156.000751289 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.699434 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:42 crc kubenswrapper[4726]: E1001 16:01:42.699620 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:43.199595033 +0000 UTC m=+156.101147610 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.699687 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:42 crc kubenswrapper[4726]: E1001 16:01:42.700020 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:43.200011764 +0000 UTC m=+156.101564341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.703819 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.704497 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.713447 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.714079 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.725841 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.801280 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.801427 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/639bf359-f577-41ad-bc87-463ac36b2e16-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"639bf359-f577-41ad-bc87-463ac36b2e16\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 16:01:42 crc kubenswrapper[4726]: E1001 16:01:42.801474 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:43.301443083 +0000 UTC m=+156.202995660 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.801532 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.801642 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/639bf359-f577-41ad-bc87-463ac36b2e16-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"639bf359-f577-41ad-bc87-463ac36b2e16\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 16:01:42 crc kubenswrapper[4726]: E1001 16:01:42.802000 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:43.301989518 +0000 UTC m=+156.203542095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.833064 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j96mp" event={"ID":"635021dd-a4fb-4bb2-ba4b-bbce8b29856f","Type":"ContainerStarted","Data":"35f3f0c888805f170460a126c386e77c2cbe275debdc34ece001686058ddb9bf"} Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.835008 4726 generic.go:334] "Generic (PLEG): container finished" podID="e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f" containerID="2a69b46eb9627f2fdd3e07e1fad741887fef5f69d93aaf6c7068b17189251583" exitCode=0 Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.835126 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" event={"ID":"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f","Type":"ContainerDied","Data":"2a69b46eb9627f2fdd3e07e1fad741887fef5f69d93aaf6c7068b17189251583"} Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.835155 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" event={"ID":"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f","Type":"ContainerStarted","Data":"28ad6b8f3b446fb0dde84c11e55f5e5fe2343624b55da92bd8741aa0c82a71b0"} Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.835165 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" event={"ID":"e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f","Type":"ContainerStarted","Data":"ef855fd06c772097952f8d73793c9867a81e9ecf9d42ce6eb11e5ae1c9f5effd"} Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.836842 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-l6d7x" event={"ID":"4739e7eb-ea30-427f-a701-29eb99ddd136","Type":"ContainerStarted","Data":"585c348f995a7b908799581b27e0045cfd26863660236c05d5ec41bab4f1323e"} Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.837716 4726 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9qvwb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.837767 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" podUID="c442dae3-d66d-47a2-b1dc-30332e947203" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.841100 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w6hhh" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.851495 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7tfz" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.857540 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" podStartSLOduration=132.857524726 podStartE2EDuration="2m12.857524726s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:42.855166813 +0000 UTC m=+155.756719400" watchObservedRunningTime="2025-10-01 16:01:42.857524726 +0000 UTC m=+155.759077303" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.902570 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.902994 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/639bf359-f577-41ad-bc87-463ac36b2e16-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"639bf359-f577-41ad-bc87-463ac36b2e16\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.903127 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/639bf359-f577-41ad-bc87-463ac36b2e16-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"639bf359-f577-41ad-bc87-463ac36b2e16\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 16:01:42 crc kubenswrapper[4726]: E1001 16:01:42.903529 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:43.403510519 +0000 UTC m=+156.305063096 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.903688 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/639bf359-f577-41ad-bc87-463ac36b2e16-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"639bf359-f577-41ad-bc87-463ac36b2e16\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.942142 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/639bf359-f577-41ad-bc87-463ac36b2e16-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"639bf359-f577-41ad-bc87-463ac36b2e16\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.966428 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.966876 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.972222 4726 patch_prober.go:28] interesting pod/apiserver-76f77b778f-vxgvr container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 01 16:01:42 crc kubenswrapper[4726]: I1001 16:01:42.972293 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" podUID="e80aae14-4c0d-45f7-ab8e-3c23dc5ee19f" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.007194 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:43 crc kubenswrapper[4726]: E1001 16:01:43.008920 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:43.508904124 +0000 UTC m=+156.410456701 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.029833 4726 patch_prober.go:28] interesting pod/router-default-5444994796-9x846 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 16:01:43 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 01 16:01:43 crc kubenswrapper[4726]: [+]process-running ok Oct 01 16:01:43 crc kubenswrapper[4726]: healthz check failed Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.029896 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9x846" podUID="60142257-8641-4fba-83ca-bb1af9f68c0d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.033144 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.108330 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:43 crc kubenswrapper[4726]: E1001 16:01:43.108620 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:43.608605157 +0000 UTC m=+156.510157724 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.211784 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:43 crc kubenswrapper[4726]: E1001 16:01:43.212365 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:43.712353327 +0000 UTC m=+156.613905904 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.277787 4726 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.312779 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:43 crc kubenswrapper[4726]: E1001 16:01:43.312917 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:43.812889872 +0000 UTC m=+156.714442459 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.313042 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:43 crc kubenswrapper[4726]: E1001 16:01:43.313378 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:43.813370495 +0000 UTC m=+156.714923072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.401562 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sm6z" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.414097 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:43 crc kubenswrapper[4726]: E1001 16:01:43.414246 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:43.914224029 +0000 UTC m=+156.815776606 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.414301 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:43 crc kubenswrapper[4726]: E1001 16:01:43.414682 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:43.914670931 +0000 UTC m=+156.816223508 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.431205 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg79v" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.516022 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:43 crc kubenswrapper[4726]: E1001 16:01:43.516445 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:44.016424957 +0000 UTC m=+156.917977534 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.587707 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.598419 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hhbb4"] Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.615321 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.617564 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:43 crc kubenswrapper[4726]: E1001 16:01:43.618082 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:44.118069752 +0000 UTC m=+157.019622329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.630508 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.631834 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hhbb4"] Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.719372 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:43 crc kubenswrapper[4726]: E1001 16:01:43.719622 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:44.219581393 +0000 UTC m=+157.121133990 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.719767 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntz7h\" (UniqueName: \"kubernetes.io/projected/6406156d-358e-4c24-9e26-921d117dc024-kube-api-access-ntz7h\") pod \"certified-operators-hhbb4\" (UID: \"6406156d-358e-4c24-9e26-921d117dc024\") " pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.719839 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6406156d-358e-4c24-9e26-921d117dc024-utilities\") pod \"certified-operators-hhbb4\" (UID: \"6406156d-358e-4c24-9e26-921d117dc024\") " pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.719880 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6406156d-358e-4c24-9e26-921d117dc024-catalog-content\") pod \"certified-operators-hhbb4\" (UID: \"6406156d-358e-4c24-9e26-921d117dc024\") " pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.719924 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:43 crc kubenswrapper[4726]: E1001 16:01:43.720397 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:44.220379954 +0000 UTC m=+157.121932531 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.788899 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gp78p"] Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.790494 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.794456 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.807787 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gp78p"] Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.821490 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.821795 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntz7h\" (UniqueName: \"kubernetes.io/projected/6406156d-358e-4c24-9e26-921d117dc024-kube-api-access-ntz7h\") pod \"certified-operators-hhbb4\" (UID: \"6406156d-358e-4c24-9e26-921d117dc024\") " pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.821838 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6406156d-358e-4c24-9e26-921d117dc024-utilities\") pod \"certified-operators-hhbb4\" (UID: \"6406156d-358e-4c24-9e26-921d117dc024\") " pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.821865 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6406156d-358e-4c24-9e26-921d117dc024-catalog-content\") pod \"certified-operators-hhbb4\" (UID: \"6406156d-358e-4c24-9e26-921d117dc024\") " pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.822803 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6406156d-358e-4c24-9e26-921d117dc024-catalog-content\") pod \"certified-operators-hhbb4\" (UID: \"6406156d-358e-4c24-9e26-921d117dc024\") " pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:01:43 crc kubenswrapper[4726]: E1001 16:01:43.822868 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 16:01:44.322853261 +0000 UTC m=+157.224405838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.823408 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6406156d-358e-4c24-9e26-921d117dc024-utilities\") pod \"certified-operators-hhbb4\" (UID: \"6406156d-358e-4c24-9e26-921d117dc024\") " pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.874212 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j96mp" event={"ID":"635021dd-a4fb-4bb2-ba4b-bbce8b29856f","Type":"ContainerStarted","Data":"6e598ca39e6eb04bb9cc3f5eeca9d53733b04bf2850ed7664bf2e38f4937e84f"} Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.874257 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j96mp" event={"ID":"635021dd-a4fb-4bb2-ba4b-bbce8b29856f","Type":"ContainerStarted","Data":"6d2c3251248be5312113fd2bd1413f2ce5371f2603640d6bc3a51bb4168fb5bf"} Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.874268 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j96mp" event={"ID":"635021dd-a4fb-4bb2-ba4b-bbce8b29856f","Type":"ContainerStarted","Data":"069b7051126d9d43274618b2328487e53b828d50e9801d789e3e3470680a3d5e"} Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.876873 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntz7h\" (UniqueName: \"kubernetes.io/projected/6406156d-358e-4c24-9e26-921d117dc024-kube-api-access-ntz7h\") pod \"certified-operators-hhbb4\" (UID: \"6406156d-358e-4c24-9e26-921d117dc024\") " pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.887317 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"639bf359-f577-41ad-bc87-463ac36b2e16","Type":"ContainerStarted","Data":"37e67bed00914548e79860cead0d78019bc69f47ba078511b14f2c23c0231f72"} Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.890857 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.903855 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-j96mp" podStartSLOduration=9.903832306 podStartE2EDuration="9.903832306s" podCreationTimestamp="2025-10-01 16:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:43.90022739 +0000 UTC m=+156.801779967" watchObservedRunningTime="2025-10-01 16:01:43.903832306 +0000 UTC m=+156.805384883" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.923678 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.923737 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g482\" (UniqueName: \"kubernetes.io/projected/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-kube-api-access-6g482\") pod \"community-operators-gp78p\" (UID: \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\") " pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.923774 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-catalog-content\") pod \"community-operators-gp78p\" (UID: \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\") " pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.923801 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-utilities\") pod \"community-operators-gp78p\" (UID: \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\") " pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:01:43 crc kubenswrapper[4726]: E1001 16:01:43.924118 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 16:01:44.424105025 +0000 UTC m=+157.325657602 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j9fr5" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.978059 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.979621 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8rfm7"] Oct 01 16:01:43 crc kubenswrapper[4726]: I1001 16:01:43.980791 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.001087 4726 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-01T16:01:43.277811489Z","Handler":null,"Name":""} Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.005942 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8rfm7"] Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.009605 4726 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.010581 4726 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.026594 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.027125 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g482\" (UniqueName: \"kubernetes.io/projected/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-kube-api-access-6g482\") pod \"community-operators-gp78p\" (UID: \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\") " pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.027323 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-catalog-content\") pod \"community-operators-gp78p\" (UID: \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\") " pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.027444 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-utilities\") pod \"community-operators-gp78p\" (UID: \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\") " pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.027623 4726 patch_prober.go:28] interesting pod/router-default-5444994796-9x846 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 16:01:44 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 01 16:01:44 crc kubenswrapper[4726]: [+]process-running ok Oct 01 16:01:44 crc kubenswrapper[4726]: healthz check failed Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.027678 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9x846" podUID="60142257-8641-4fba-83ca-bb1af9f68c0d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.030776 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-catalog-content\") pod \"community-operators-gp78p\" (UID: \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\") " pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.031408 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-utilities\") pod \"community-operators-gp78p\" (UID: \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\") " pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.035472 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.047428 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g482\" (UniqueName: \"kubernetes.io/projected/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-kube-api-access-6g482\") pod \"community-operators-gp78p\" (UID: \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\") " pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.109101 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.131680 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.131739 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b6165c-ee06-406a-b01e-35817eab4a32-catalog-content\") pod \"certified-operators-8rfm7\" (UID: \"30b6165c-ee06-406a-b01e-35817eab4a32\") " pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.131779 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbqsm\" (UniqueName: \"kubernetes.io/projected/30b6165c-ee06-406a-b01e-35817eab4a32-kube-api-access-zbqsm\") pod \"certified-operators-8rfm7\" (UID: \"30b6165c-ee06-406a-b01e-35817eab4a32\") " pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.131839 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b6165c-ee06-406a-b01e-35817eab4a32-utilities\") pod \"certified-operators-8rfm7\" (UID: \"30b6165c-ee06-406a-b01e-35817eab4a32\") " pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.142191 4726 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.142221 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.180560 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4qzjh"] Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.183511 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.187479 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4qzjh"] Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.196349 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j9fr5\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.208407 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hhbb4"] Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.232635 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b6165c-ee06-406a-b01e-35817eab4a32-catalog-content\") pod \"certified-operators-8rfm7\" (UID: \"30b6165c-ee06-406a-b01e-35817eab4a32\") " pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.232683 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbqsm\" (UniqueName: \"kubernetes.io/projected/30b6165c-ee06-406a-b01e-35817eab4a32-kube-api-access-zbqsm\") pod \"certified-operators-8rfm7\" (UID: \"30b6165c-ee06-406a-b01e-35817eab4a32\") " pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.232775 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b6165c-ee06-406a-b01e-35817eab4a32-utilities\") pod \"certified-operators-8rfm7\" (UID: \"30b6165c-ee06-406a-b01e-35817eab4a32\") " pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.233692 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b6165c-ee06-406a-b01e-35817eab4a32-catalog-content\") pod \"certified-operators-8rfm7\" (UID: \"30b6165c-ee06-406a-b01e-35817eab4a32\") " pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.233730 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b6165c-ee06-406a-b01e-35817eab4a32-utilities\") pod \"certified-operators-8rfm7\" (UID: \"30b6165c-ee06-406a-b01e-35817eab4a32\") " pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.251080 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbqsm\" (UniqueName: \"kubernetes.io/projected/30b6165c-ee06-406a-b01e-35817eab4a32-kube-api-access-zbqsm\") pod \"certified-operators-8rfm7\" (UID: \"30b6165c-ee06-406a-b01e-35817eab4a32\") " pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.310623 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.334878 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzxq8\" (UniqueName: \"kubernetes.io/projected/03392899-81fb-49bf-8f23-8d056c3a8e19-kube-api-access-rzxq8\") pod \"community-operators-4qzjh\" (UID: \"03392899-81fb-49bf-8f23-8d056c3a8e19\") " pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.335301 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03392899-81fb-49bf-8f23-8d056c3a8e19-catalog-content\") pod \"community-operators-4qzjh\" (UID: \"03392899-81fb-49bf-8f23-8d056c3a8e19\") " pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.335354 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03392899-81fb-49bf-8f23-8d056c3a8e19-utilities\") pod \"community-operators-4qzjh\" (UID: \"03392899-81fb-49bf-8f23-8d056c3a8e19\") " pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.335870 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gp78p"] Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.437339 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzxq8\" (UniqueName: \"kubernetes.io/projected/03392899-81fb-49bf-8f23-8d056c3a8e19-kube-api-access-rzxq8\") pod \"community-operators-4qzjh\" (UID: \"03392899-81fb-49bf-8f23-8d056c3a8e19\") " pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.437530 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03392899-81fb-49bf-8f23-8d056c3a8e19-catalog-content\") pod \"community-operators-4qzjh\" (UID: \"03392899-81fb-49bf-8f23-8d056c3a8e19\") " pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.437587 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03392899-81fb-49bf-8f23-8d056c3a8e19-utilities\") pod \"community-operators-4qzjh\" (UID: \"03392899-81fb-49bf-8f23-8d056c3a8e19\") " pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.438148 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03392899-81fb-49bf-8f23-8d056c3a8e19-catalog-content\") pod \"community-operators-4qzjh\" (UID: \"03392899-81fb-49bf-8f23-8d056c3a8e19\") " pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.438882 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03392899-81fb-49bf-8f23-8d056c3a8e19-utilities\") pod \"community-operators-4qzjh\" (UID: \"03392899-81fb-49bf-8f23-8d056c3a8e19\") " pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.457285 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzxq8\" (UniqueName: \"kubernetes.io/projected/03392899-81fb-49bf-8f23-8d056c3a8e19-kube-api-access-rzxq8\") pod \"community-operators-4qzjh\" (UID: \"03392899-81fb-49bf-8f23-8d056c3a8e19\") " pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.496654 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.526406 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.680136 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8rfm7"] Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.896513 4726 generic.go:334] "Generic (PLEG): container finished" podID="639bf359-f577-41ad-bc87-463ac36b2e16" containerID="ed1ebdf5169f90d7ad6d45d6b1472ad4b72173bc906b551471989401c832d087" exitCode=0 Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.896587 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"639bf359-f577-41ad-bc87-463ac36b2e16","Type":"ContainerDied","Data":"ed1ebdf5169f90d7ad6d45d6b1472ad4b72173bc906b551471989401c832d087"} Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.898553 4726 generic.go:334] "Generic (PLEG): container finished" podID="6406156d-358e-4c24-9e26-921d117dc024" containerID="f8889a56d11176b7ea0555ddcbe3eeafb4287e5c8c3dc141716232345a3a4ec7" exitCode=0 Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.898622 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hhbb4" event={"ID":"6406156d-358e-4c24-9e26-921d117dc024","Type":"ContainerDied","Data":"f8889a56d11176b7ea0555ddcbe3eeafb4287e5c8c3dc141716232345a3a4ec7"} Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.898649 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hhbb4" event={"ID":"6406156d-358e-4c24-9e26-921d117dc024","Type":"ContainerStarted","Data":"38098a5c023d48324aa8e5a4b494b2576d34934411107b77105c54166c920a14"} Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.904543 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rfm7" event={"ID":"30b6165c-ee06-406a-b01e-35817eab4a32","Type":"ContainerStarted","Data":"92b214e929c44e968c8e4f124511c320fc1e5cb086957e4f50d71d603f914508"} Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.904623 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.906745 4726 generic.go:334] "Generic (PLEG): container finished" podID="6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" containerID="8a8825e1ae9ee2b2363e5d02ad358f197b868264660b6da6403a18355b6cf3d3" exitCode=0 Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.906874 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gp78p" event={"ID":"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb","Type":"ContainerDied","Data":"8a8825e1ae9ee2b2363e5d02ad358f197b868264660b6da6403a18355b6cf3d3"} Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.906906 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gp78p" event={"ID":"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb","Type":"ContainerStarted","Data":"571b5db544d7c381d35c4249c1dbe752165549ec94063bc9b7e86f23e1b0732f"} Oct 01 16:01:44 crc kubenswrapper[4726]: I1001 16:01:44.908802 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4qzjh"] Oct 01 16:01:44 crc kubenswrapper[4726]: W1001 16:01:44.925289 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03392899_81fb_49bf_8f23_8d056c3a8e19.slice/crio-a3818d3414ab65ad0a5331b4e58c7873a25ddd17b9bd1b00d2968244d52ecbef WatchSource:0}: Error finding container a3818d3414ab65ad0a5331b4e58c7873a25ddd17b9bd1b00d2968244d52ecbef: Status 404 returned error can't find the container with id a3818d3414ab65ad0a5331b4e58c7873a25ddd17b9bd1b00d2968244d52ecbef Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.022894 4726 patch_prober.go:28] interesting pod/router-default-5444994796-9x846 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 16:01:45 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 01 16:01:45 crc kubenswrapper[4726]: [+]process-running ok Oct 01 16:01:45 crc kubenswrapper[4726]: healthz check failed Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.022958 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9x846" podUID="60142257-8641-4fba-83ca-bb1af9f68c0d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.045618 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j9fr5"] Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.573365 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9m52k"] Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.575393 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.577654 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.587636 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9m52k"] Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.662193 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b53fe26-735b-47b1-b772-8de0898560ee-utilities\") pod \"redhat-marketplace-9m52k\" (UID: \"9b53fe26-735b-47b1-b772-8de0898560ee\") " pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.662288 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b53fe26-735b-47b1-b772-8de0898560ee-catalog-content\") pod \"redhat-marketplace-9m52k\" (UID: \"9b53fe26-735b-47b1-b772-8de0898560ee\") " pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.662463 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6b9s\" (UniqueName: \"kubernetes.io/projected/9b53fe26-735b-47b1-b772-8de0898560ee-kube-api-access-n6b9s\") pod \"redhat-marketplace-9m52k\" (UID: \"9b53fe26-735b-47b1-b772-8de0898560ee\") " pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.763906 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b53fe26-735b-47b1-b772-8de0898560ee-utilities\") pod \"redhat-marketplace-9m52k\" (UID: \"9b53fe26-735b-47b1-b772-8de0898560ee\") " pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.763961 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b53fe26-735b-47b1-b772-8de0898560ee-catalog-content\") pod \"redhat-marketplace-9m52k\" (UID: \"9b53fe26-735b-47b1-b772-8de0898560ee\") " pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.764005 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6b9s\" (UniqueName: \"kubernetes.io/projected/9b53fe26-735b-47b1-b772-8de0898560ee-kube-api-access-n6b9s\") pod \"redhat-marketplace-9m52k\" (UID: \"9b53fe26-735b-47b1-b772-8de0898560ee\") " pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.764645 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b53fe26-735b-47b1-b772-8de0898560ee-catalog-content\") pod \"redhat-marketplace-9m52k\" (UID: \"9b53fe26-735b-47b1-b772-8de0898560ee\") " pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.764657 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b53fe26-735b-47b1-b772-8de0898560ee-utilities\") pod \"redhat-marketplace-9m52k\" (UID: \"9b53fe26-735b-47b1-b772-8de0898560ee\") " pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.784862 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6b9s\" (UniqueName: \"kubernetes.io/projected/9b53fe26-735b-47b1-b772-8de0898560ee-kube-api-access-n6b9s\") pod \"redhat-marketplace-9m52k\" (UID: \"9b53fe26-735b-47b1-b772-8de0898560ee\") " pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.813709 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.918592 4726 generic.go:334] "Generic (PLEG): container finished" podID="30b6165c-ee06-406a-b01e-35817eab4a32" containerID="7d6128006e57892a941cfc214d3dc9d0db634067bddcca8cff51d3f3fa1db45c" exitCode=0 Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.918725 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rfm7" event={"ID":"30b6165c-ee06-406a-b01e-35817eab4a32","Type":"ContainerDied","Data":"7d6128006e57892a941cfc214d3dc9d0db634067bddcca8cff51d3f3fa1db45c"} Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.920942 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" event={"ID":"496e7e53-b6f6-48b1-9657-a091d21c301f","Type":"ContainerStarted","Data":"944ceea21ccc4b37f40718a1146700fe4fb37c6baf6ff823552f3b984b295869"} Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.920990 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" event={"ID":"496e7e53-b6f6-48b1-9657-a091d21c301f","Type":"ContainerStarted","Data":"19f64afe7167ee0df00b74e33c69ddc2b79077ede2f6bf7de0cd929487b9d597"} Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.921083 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.923253 4726 generic.go:334] "Generic (PLEG): container finished" podID="03392899-81fb-49bf-8f23-8d056c3a8e19" containerID="308568c843e938acde8fe2bcb6e0e659cba7fd1db0e2fe88c19d626fc2ee2e99" exitCode=0 Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.923351 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qzjh" event={"ID":"03392899-81fb-49bf-8f23-8d056c3a8e19","Type":"ContainerDied","Data":"308568c843e938acde8fe2bcb6e0e659cba7fd1db0e2fe88c19d626fc2ee2e99"} Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.923388 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qzjh" event={"ID":"03392899-81fb-49bf-8f23-8d056c3a8e19","Type":"ContainerStarted","Data":"a3818d3414ab65ad0a5331b4e58c7873a25ddd17b9bd1b00d2968244d52ecbef"} Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.928450 4726 generic.go:334] "Generic (PLEG): container finished" podID="f8cbbf14-6dac-425e-9cb6-8579f7a150ee" containerID="821ee00b78fb3620fef1e00d43d963d218a9f1d739df5183f5fd83421f603b7e" exitCode=0 Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.928755 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" event={"ID":"f8cbbf14-6dac-425e-9cb6-8579f7a150ee","Type":"ContainerDied","Data":"821ee00b78fb3620fef1e00d43d963d218a9f1d739df5183f5fd83421f603b7e"} Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.939783 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.972561 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" podStartSLOduration=135.972538132 podStartE2EDuration="2m15.972538132s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:45.968795762 +0000 UTC m=+158.870348379" watchObservedRunningTime="2025-10-01 16:01:45.972538132 +0000 UTC m=+158.874090709" Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.982393 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g6v9s"] Oct 01 16:01:45 crc kubenswrapper[4726]: I1001 16:01:45.990316 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.010280 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6v9s"] Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.022449 4726 patch_prober.go:28] interesting pod/router-default-5444994796-9x846 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 16:01:46 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 01 16:01:46 crc kubenswrapper[4726]: [+]process-running ok Oct 01 16:01:46 crc kubenswrapper[4726]: healthz check failed Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.022507 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9x846" podUID="60142257-8641-4fba-83ca-bb1af9f68c0d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.171983 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcddw\" (UniqueName: \"kubernetes.io/projected/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-kube-api-access-vcddw\") pod \"redhat-marketplace-g6v9s\" (UID: \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\") " pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.172403 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-utilities\") pod \"redhat-marketplace-g6v9s\" (UID: \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\") " pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.172442 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-catalog-content\") pod \"redhat-marketplace-g6v9s\" (UID: \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\") " pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.180320 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.243620 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9m52k"] Oct 01 16:01:46 crc kubenswrapper[4726]: W1001 16:01:46.259791 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b53fe26_735b_47b1_b772_8de0898560ee.slice/crio-cb22d203e077bb599e6b97969e2a3d5df12c11b62ae2f0adaa67d97c4a9252fa WatchSource:0}: Error finding container cb22d203e077bb599e6b97969e2a3d5df12c11b62ae2f0adaa67d97c4a9252fa: Status 404 returned error can't find the container with id cb22d203e077bb599e6b97969e2a3d5df12c11b62ae2f0adaa67d97c4a9252fa Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.273156 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/639bf359-f577-41ad-bc87-463ac36b2e16-kubelet-dir\") pod \"639bf359-f577-41ad-bc87-463ac36b2e16\" (UID: \"639bf359-f577-41ad-bc87-463ac36b2e16\") " Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.273317 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/639bf359-f577-41ad-bc87-463ac36b2e16-kube-api-access\") pod \"639bf359-f577-41ad-bc87-463ac36b2e16\" (UID: \"639bf359-f577-41ad-bc87-463ac36b2e16\") " Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.273344 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/639bf359-f577-41ad-bc87-463ac36b2e16-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "639bf359-f577-41ad-bc87-463ac36b2e16" (UID: "639bf359-f577-41ad-bc87-463ac36b2e16"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.273558 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-utilities\") pod \"redhat-marketplace-g6v9s\" (UID: \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\") " pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.273596 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-catalog-content\") pod \"redhat-marketplace-g6v9s\" (UID: \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\") " pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.273683 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcddw\" (UniqueName: \"kubernetes.io/projected/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-kube-api-access-vcddw\") pod \"redhat-marketplace-g6v9s\" (UID: \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\") " pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.273736 4726 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/639bf359-f577-41ad-bc87-463ac36b2e16-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.274003 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-utilities\") pod \"redhat-marketplace-g6v9s\" (UID: \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\") " pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.274140 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-catalog-content\") pod \"redhat-marketplace-g6v9s\" (UID: \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\") " pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.281575 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/639bf359-f577-41ad-bc87-463ac36b2e16-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "639bf359-f577-41ad-bc87-463ac36b2e16" (UID: "639bf359-f577-41ad-bc87-463ac36b2e16"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.293522 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcddw\" (UniqueName: \"kubernetes.io/projected/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-kube-api-access-vcddw\") pod \"redhat-marketplace-g6v9s\" (UID: \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\") " pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.330576 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.377349 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/639bf359-f577-41ad-bc87-463ac36b2e16-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.540370 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6v9s"] Oct 01 16:01:46 crc kubenswrapper[4726]: W1001 16:01:46.580349 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ecefa98_2ab6_4e4c_8a0b_589d845568f1.slice/crio-2d71ab888087b0e8e6a5772bdc8468d17974f5506d6ffffce8258fb9ea64f9f8 WatchSource:0}: Error finding container 2d71ab888087b0e8e6a5772bdc8468d17974f5506d6ffffce8258fb9ea64f9f8: Status 404 returned error can't find the container with id 2d71ab888087b0e8e6a5772bdc8468d17974f5506d6ffffce8258fb9ea64f9f8 Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.642505 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.950939 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.952815 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"639bf359-f577-41ad-bc87-463ac36b2e16","Type":"ContainerDied","Data":"37e67bed00914548e79860cead0d78019bc69f47ba078511b14f2c23c0231f72"} Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.952869 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37e67bed00914548e79860cead0d78019bc69f47ba078511b14f2c23c0231f72" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.955040 4726 generic.go:334] "Generic (PLEG): container finished" podID="8ecefa98-2ab6-4e4c-8a0b-589d845568f1" containerID="669f5bf0a72e8665d8dd1dfc868c8db666bfe3d5fd8230273793fb982e95333d" exitCode=0 Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.955141 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6v9s" event={"ID":"8ecefa98-2ab6-4e4c-8a0b-589d845568f1","Type":"ContainerDied","Data":"669f5bf0a72e8665d8dd1dfc868c8db666bfe3d5fd8230273793fb982e95333d"} Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.955403 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6v9s" event={"ID":"8ecefa98-2ab6-4e4c-8a0b-589d845568f1","Type":"ContainerStarted","Data":"2d71ab888087b0e8e6a5772bdc8468d17974f5506d6ffffce8258fb9ea64f9f8"} Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.959341 4726 generic.go:334] "Generic (PLEG): container finished" podID="9b53fe26-735b-47b1-b772-8de0898560ee" containerID="c477b1235eeba774a5ea24f37c4bfb8c59e5618f8784fadf2962617e71a3618c" exitCode=0 Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.960446 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9m52k" event={"ID":"9b53fe26-735b-47b1-b772-8de0898560ee","Type":"ContainerDied","Data":"c477b1235eeba774a5ea24f37c4bfb8c59e5618f8784fadf2962617e71a3618c"} Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.960476 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9m52k" event={"ID":"9b53fe26-735b-47b1-b772-8de0898560ee","Type":"ContainerStarted","Data":"cb22d203e077bb599e6b97969e2a3d5df12c11b62ae2f0adaa67d97c4a9252fa"} Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.982825 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dk98l"] Oct 01 16:01:46 crc kubenswrapper[4726]: E1001 16:01:46.983136 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="639bf359-f577-41ad-bc87-463ac36b2e16" containerName="pruner" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.983172 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="639bf359-f577-41ad-bc87-463ac36b2e16" containerName="pruner" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.983278 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="639bf359-f577-41ad-bc87-463ac36b2e16" containerName="pruner" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.984816 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.986542 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 01 16:01:46 crc kubenswrapper[4726]: I1001 16:01:46.996641 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dk98l"] Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.020027 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.027538 4726 patch_prober.go:28] interesting pod/router-default-5444994796-9x846 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 16:01:47 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 01 16:01:47 crc kubenswrapper[4726]: [+]process-running ok Oct 01 16:01:47 crc kubenswrapper[4726]: healthz check failed Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.027604 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9x846" podUID="60142257-8641-4fba-83ca-bb1af9f68c0d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.090454 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-catalog-content\") pod \"redhat-operators-dk98l\" (UID: \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\") " pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.096464 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncf6z\" (UniqueName: \"kubernetes.io/projected/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-kube-api-access-ncf6z\") pod \"redhat-operators-dk98l\" (UID: \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\") " pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.096588 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-utilities\") pod \"redhat-operators-dk98l\" (UID: \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\") " pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.195218 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.195280 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.197260 4726 patch_prober.go:28] interesting pod/console-f9d7485db-w4c9z container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.197382 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-w4c9z" podUID="79d896b5-b34d-4206-92d8-99caeafdffb2" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.197702 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncf6z\" (UniqueName: \"kubernetes.io/projected/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-kube-api-access-ncf6z\") pod \"redhat-operators-dk98l\" (UID: \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\") " pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.198487 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-utilities\") pod \"redhat-operators-dk98l\" (UID: \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\") " pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.199088 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-utilities\") pod \"redhat-operators-dk98l\" (UID: \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\") " pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.199249 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-catalog-content\") pod \"redhat-operators-dk98l\" (UID: \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\") " pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.199581 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-catalog-content\") pod \"redhat-operators-dk98l\" (UID: \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\") " pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.216859 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncf6z\" (UniqueName: \"kubernetes.io/projected/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-kube-api-access-ncf6z\") pod \"redhat-operators-dk98l\" (UID: \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\") " pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.216973 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.220169 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-5tjmt container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.220216 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-5tjmt" podUID="b8cf6941-f833-4465-a3e9-9efea4b9b5b0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.224417 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-5tjmt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.224471 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5tjmt" podUID="b8cf6941-f833-4465-a3e9-9efea4b9b5b0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.300738 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqkk7\" (UniqueName: \"kubernetes.io/projected/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-kube-api-access-mqkk7\") pod \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\" (UID: \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\") " Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.300814 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-secret-volume\") pod \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\" (UID: \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\") " Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.300953 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-config-volume\") pod \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\" (UID: \"f8cbbf14-6dac-425e-9cb6-8579f7a150ee\") " Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.302504 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-config-volume" (OuterVolumeSpecName: "config-volume") pod "f8cbbf14-6dac-425e-9cb6-8579f7a150ee" (UID: "f8cbbf14-6dac-425e-9cb6-8579f7a150ee"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.306276 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f8cbbf14-6dac-425e-9cb6-8579f7a150ee" (UID: "f8cbbf14-6dac-425e-9cb6-8579f7a150ee"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.306902 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-kube-api-access-mqkk7" (OuterVolumeSpecName: "kube-api-access-mqkk7") pod "f8cbbf14-6dac-425e-9cb6-8579f7a150ee" (UID: "f8cbbf14-6dac-425e-9cb6-8579f7a150ee"). InnerVolumeSpecName "kube-api-access-mqkk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.349827 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.381260 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nw2vt"] Oct 01 16:01:47 crc kubenswrapper[4726]: E1001 16:01:47.384978 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8cbbf14-6dac-425e-9cb6-8579f7a150ee" containerName="collect-profiles" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.385017 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8cbbf14-6dac-425e-9cb6-8579f7a150ee" containerName="collect-profiles" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.385221 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8cbbf14-6dac-425e-9cb6-8579f7a150ee" containerName="collect-profiles" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.389589 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.395703 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nw2vt"] Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.408804 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.408839 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqkk7\" (UniqueName: \"kubernetes.io/projected/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-kube-api-access-mqkk7\") on node \"crc\" DevicePath \"\"" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.408850 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8cbbf14-6dac-425e-9cb6-8579f7a150ee-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.512248 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzlxp\" (UniqueName: \"kubernetes.io/projected/7502490c-e673-49e8-9645-775bdf4899df-kube-api-access-zzlxp\") pod \"redhat-operators-nw2vt\" (UID: \"7502490c-e673-49e8-9645-775bdf4899df\") " pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.512296 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7502490c-e673-49e8-9645-775bdf4899df-catalog-content\") pod \"redhat-operators-nw2vt\" (UID: \"7502490c-e673-49e8-9645-775bdf4899df\") " pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.512350 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7502490c-e673-49e8-9645-775bdf4899df-utilities\") pod \"redhat-operators-nw2vt\" (UID: \"7502490c-e673-49e8-9645-775bdf4899df\") " pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.609717 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dk98l"] Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.613763 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzlxp\" (UniqueName: \"kubernetes.io/projected/7502490c-e673-49e8-9645-775bdf4899df-kube-api-access-zzlxp\") pod \"redhat-operators-nw2vt\" (UID: \"7502490c-e673-49e8-9645-775bdf4899df\") " pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.613823 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7502490c-e673-49e8-9645-775bdf4899df-catalog-content\") pod \"redhat-operators-nw2vt\" (UID: \"7502490c-e673-49e8-9645-775bdf4899df\") " pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.613850 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7502490c-e673-49e8-9645-775bdf4899df-utilities\") pod \"redhat-operators-nw2vt\" (UID: \"7502490c-e673-49e8-9645-775bdf4899df\") " pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.614395 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7502490c-e673-49e8-9645-775bdf4899df-utilities\") pod \"redhat-operators-nw2vt\" (UID: \"7502490c-e673-49e8-9645-775bdf4899df\") " pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.614406 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7502490c-e673-49e8-9645-775bdf4899df-catalog-content\") pod \"redhat-operators-nw2vt\" (UID: \"7502490c-e673-49e8-9645-775bdf4899df\") " pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:01:47 crc kubenswrapper[4726]: W1001 16:01:47.617194 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a7a60ae_9178_467a_b52e_8e9c37dc64a3.slice/crio-b10e4c3898eb00bc4f3c6e51473710cdc784636f60a7b8e80d9e6838b03bbac2 WatchSource:0}: Error finding container b10e4c3898eb00bc4f3c6e51473710cdc784636f60a7b8e80d9e6838b03bbac2: Status 404 returned error can't find the container with id b10e4c3898eb00bc4f3c6e51473710cdc784636f60a7b8e80d9e6838b03bbac2 Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.645420 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzlxp\" (UniqueName: \"kubernetes.io/projected/7502490c-e673-49e8-9645-775bdf4899df-kube-api-access-zzlxp\") pod \"redhat-operators-nw2vt\" (UID: \"7502490c-e673-49e8-9645-775bdf4899df\") " pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.757101 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.972231 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" event={"ID":"f8cbbf14-6dac-425e-9cb6-8579f7a150ee","Type":"ContainerDied","Data":"b08a748bc61d06578024920db7aff0e4aab43fd87461db4491b0b1c0b9146eca"} Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.972765 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b08a748bc61d06578024920db7aff0e4aab43fd87461db4491b0b1c0b9146eca" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.972841 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.972382 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4" Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.975834 4726 generic.go:334] "Generic (PLEG): container finished" podID="9a7a60ae-9178-467a-b52e-8e9c37dc64a3" containerID="ae0edf0ab9f075f4124b0fa4a9107804571ba9f43860034ad1ee2e9283f275ee" exitCode=0 Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.975895 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dk98l" event={"ID":"9a7a60ae-9178-467a-b52e-8e9c37dc64a3","Type":"ContainerDied","Data":"ae0edf0ab9f075f4124b0fa4a9107804571ba9f43860034ad1ee2e9283f275ee"} Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.975922 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dk98l" event={"ID":"9a7a60ae-9178-467a-b52e-8e9c37dc64a3","Type":"ContainerStarted","Data":"b10e4c3898eb00bc4f3c6e51473710cdc784636f60a7b8e80d9e6838b03bbac2"} Oct 01 16:01:47 crc kubenswrapper[4726]: I1001 16:01:47.977802 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-vxgvr" Oct 01 16:01:48 crc kubenswrapper[4726]: I1001 16:01:48.023698 4726 patch_prober.go:28] interesting pod/router-default-5444994796-9x846 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 16:01:48 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 01 16:01:48 crc kubenswrapper[4726]: [+]process-running ok Oct 01 16:01:48 crc kubenswrapper[4726]: healthz check failed Oct 01 16:01:48 crc kubenswrapper[4726]: I1001 16:01:48.023794 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9x846" podUID="60142257-8641-4fba-83ca-bb1af9f68c0d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 16:01:48 crc kubenswrapper[4726]: I1001 16:01:48.024230 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nw2vt"] Oct 01 16:01:48 crc kubenswrapper[4726]: W1001 16:01:48.118238 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7502490c_e673_49e8_9645_775bdf4899df.slice/crio-5fd9a2bc5ee54c59f702c4f97cc87fce38f09969cc016fc5edb26d68f2c5e75a WatchSource:0}: Error finding container 5fd9a2bc5ee54c59f702c4f97cc87fce38f09969cc016fc5edb26d68f2c5e75a: Status 404 returned error can't find the container with id 5fd9a2bc5ee54c59f702c4f97cc87fce38f09969cc016fc5edb26d68f2c5e75a Oct 01 16:01:48 crc kubenswrapper[4726]: I1001 16:01:48.997257 4726 generic.go:334] "Generic (PLEG): container finished" podID="7502490c-e673-49e8-9645-775bdf4899df" containerID="8d927f5093b4a80920687a0d4b93b79daff81ee2e7255a6a55e764d380da56f7" exitCode=0 Oct 01 16:01:48 crc kubenswrapper[4726]: I1001 16:01:48.997409 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nw2vt" event={"ID":"7502490c-e673-49e8-9645-775bdf4899df","Type":"ContainerDied","Data":"8d927f5093b4a80920687a0d4b93b79daff81ee2e7255a6a55e764d380da56f7"} Oct 01 16:01:48 crc kubenswrapper[4726]: I1001 16:01:48.997643 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nw2vt" event={"ID":"7502490c-e673-49e8-9645-775bdf4899df","Type":"ContainerStarted","Data":"5fd9a2bc5ee54c59f702c4f97cc87fce38f09969cc016fc5edb26d68f2c5e75a"} Oct 01 16:01:49 crc kubenswrapper[4726]: I1001 16:01:49.037848 4726 patch_prober.go:28] interesting pod/router-default-5444994796-9x846 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 16:01:49 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 01 16:01:49 crc kubenswrapper[4726]: [+]process-running ok Oct 01 16:01:49 crc kubenswrapper[4726]: healthz check failed Oct 01 16:01:49 crc kubenswrapper[4726]: I1001 16:01:49.037929 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9x846" podUID="60142257-8641-4fba-83ca-bb1af9f68c0d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 16:01:49 crc kubenswrapper[4726]: I1001 16:01:49.441312 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 16:01:49 crc kubenswrapper[4726]: I1001 16:01:49.441970 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 16:01:49 crc kubenswrapper[4726]: I1001 16:01:49.449831 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 01 16:01:49 crc kubenswrapper[4726]: I1001 16:01:49.450238 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 01 16:01:49 crc kubenswrapper[4726]: I1001 16:01:49.459981 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 16:01:49 crc kubenswrapper[4726]: I1001 16:01:49.540834 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 16:01:49 crc kubenswrapper[4726]: I1001 16:01:49.540882 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 16:01:49 crc kubenswrapper[4726]: I1001 16:01:49.642763 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 16:01:49 crc kubenswrapper[4726]: I1001 16:01:49.642819 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 16:01:49 crc kubenswrapper[4726]: I1001 16:01:49.642882 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 16:01:49 crc kubenswrapper[4726]: I1001 16:01:49.684841 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 16:01:49 crc kubenswrapper[4726]: I1001 16:01:49.769186 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 16:01:50 crc kubenswrapper[4726]: I1001 16:01:50.023043 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:50 crc kubenswrapper[4726]: I1001 16:01:50.026456 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-9x846" Oct 01 16:01:50 crc kubenswrapper[4726]: I1001 16:01:50.217112 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 16:01:52 crc kubenswrapper[4726]: I1001 16:01:52.442162 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-l6d7x" Oct 01 16:01:53 crc kubenswrapper[4726]: I1001 16:01:53.413419 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:01:53 crc kubenswrapper[4726]: I1001 16:01:53.413792 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:01:53 crc kubenswrapper[4726]: I1001 16:01:53.513733 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs\") pod \"network-metrics-daemon-4pzd8\" (UID: \"2a22e7a8-726a-4538-a27d-44c58f89f0b1\") " pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:53 crc kubenswrapper[4726]: I1001 16:01:53.519086 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2a22e7a8-726a-4538-a27d-44c58f89f0b1-metrics-certs\") pod \"network-metrics-daemon-4pzd8\" (UID: \"2a22e7a8-726a-4538-a27d-44c58f89f0b1\") " pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:53 crc kubenswrapper[4726]: I1001 16:01:53.744891 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4pzd8" Oct 01 16:01:57 crc kubenswrapper[4726]: I1001 16:01:57.211446 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:57 crc kubenswrapper[4726]: I1001 16:01:57.214861 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:01:57 crc kubenswrapper[4726]: I1001 16:01:57.226891 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-5tjmt" Oct 01 16:02:01 crc kubenswrapper[4726]: W1001 16:02:01.292318 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod21dc5d81_ff6f_4d8d_afc6_30d6dd87f15c.slice/crio-a4e28a95a48db869d800c3f1c2a7f3dda48ecf4bc4abae6c63fe0d01b7ccb65a WatchSource:0}: Error finding container a4e28a95a48db869d800c3f1c2a7f3dda48ecf4bc4abae6c63fe0d01b7ccb65a: Status 404 returned error can't find the container with id a4e28a95a48db869d800c3f1c2a7f3dda48ecf4bc4abae6c63fe0d01b7ccb65a Oct 01 16:02:02 crc kubenswrapper[4726]: I1001 16:02:02.112370 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c","Type":"ContainerStarted","Data":"a4e28a95a48db869d800c3f1c2a7f3dda48ecf4bc4abae6c63fe0d01b7ccb65a"} Oct 01 16:02:04 crc kubenswrapper[4726]: I1001 16:02:04.503838 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:02:17 crc kubenswrapper[4726]: I1001 16:02:17.378034 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-szpqh" Oct 01 16:02:18 crc kubenswrapper[4726]: I1001 16:02:18.879528 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 16:02:23 crc kubenswrapper[4726]: I1001 16:02:23.413978 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:02:23 crc kubenswrapper[4726]: I1001 16:02:23.414110 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:02:23 crc kubenswrapper[4726]: E1001 16:02:23.681483 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 01 16:02:23 crc kubenswrapper[4726]: E1001 16:02:23.682003 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rzxq8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-4qzjh_openshift-marketplace(03392899-81fb-49bf-8f23-8d056c3a8e19): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 16:02:23 crc kubenswrapper[4726]: E1001 16:02:23.685163 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-4qzjh" podUID="03392899-81fb-49bf-8f23-8d056c3a8e19" Oct 01 16:02:26 crc kubenswrapper[4726]: E1001 16:02:26.015940 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-4qzjh" podUID="03392899-81fb-49bf-8f23-8d056c3a8e19" Oct 01 16:02:26 crc kubenswrapper[4726]: E1001 16:02:26.173600 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 01 16:02:26 crc kubenswrapper[4726]: E1001 16:02:26.173766 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ntz7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-hhbb4_openshift-marketplace(6406156d-358e-4c24-9e26-921d117dc024): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 16:02:26 crc kubenswrapper[4726]: E1001 16:02:26.174968 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-hhbb4" podUID="6406156d-358e-4c24-9e26-921d117dc024" Oct 01 16:02:30 crc kubenswrapper[4726]: E1001 16:02:30.689301 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-hhbb4" podUID="6406156d-358e-4c24-9e26-921d117dc024" Oct 01 16:02:31 crc kubenswrapper[4726]: E1001 16:02:31.102948 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 01 16:02:31 crc kubenswrapper[4726]: E1001 16:02:31.103300 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n6b9s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-9m52k_openshift-marketplace(9b53fe26-735b-47b1-b772-8de0898560ee): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 16:02:31 crc kubenswrapper[4726]: E1001 16:02:31.105835 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-9m52k" podUID="9b53fe26-735b-47b1-b772-8de0898560ee" Oct 01 16:02:31 crc kubenswrapper[4726]: I1001 16:02:31.128184 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4pzd8"] Oct 01 16:02:31 crc kubenswrapper[4726]: E1001 16:02:31.133194 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 01 16:02:31 crc kubenswrapper[4726]: E1001 16:02:31.133343 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vcddw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-g6v9s_openshift-marketplace(8ecefa98-2ab6-4e4c-8a0b-589d845568f1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 16:02:31 crc kubenswrapper[4726]: E1001 16:02:31.134621 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-g6v9s" podUID="8ecefa98-2ab6-4e4c-8a0b-589d845568f1" Oct 01 16:02:31 crc kubenswrapper[4726]: E1001 16:02:31.214367 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 01 16:02:31 crc kubenswrapper[4726]: E1001 16:02:31.214550 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zbqsm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-8rfm7_openshift-marketplace(30b6165c-ee06-406a-b01e-35817eab4a32): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 16:02:31 crc kubenswrapper[4726]: E1001 16:02:31.215742 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-8rfm7" podUID="30b6165c-ee06-406a-b01e-35817eab4a32" Oct 01 16:02:31 crc kubenswrapper[4726]: E1001 16:02:31.403778 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-g6v9s" podUID="8ecefa98-2ab6-4e4c-8a0b-589d845568f1" Oct 01 16:02:31 crc kubenswrapper[4726]: E1001 16:02:31.403871 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9m52k" podUID="9b53fe26-735b-47b1-b772-8de0898560ee" Oct 01 16:02:31 crc kubenswrapper[4726]: E1001 16:02:31.403915 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-8rfm7" podUID="30b6165c-ee06-406a-b01e-35817eab4a32" Oct 01 16:02:31 crc kubenswrapper[4726]: W1001 16:02:31.413184 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a22e7a8_726a_4538_a27d_44c58f89f0b1.slice/crio-bf9a933adb8b2b12301e0767d0d62e790534ca6661f7532133f934e93c2930fb WatchSource:0}: Error finding container bf9a933adb8b2b12301e0767d0d62e790534ca6661f7532133f934e93c2930fb: Status 404 returned error can't find the container with id bf9a933adb8b2b12301e0767d0d62e790534ca6661f7532133f934e93c2930fb Oct 01 16:02:32 crc kubenswrapper[4726]: I1001 16:02:32.301234 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" event={"ID":"2a22e7a8-726a-4538-a27d-44c58f89f0b1","Type":"ContainerStarted","Data":"bf9a933adb8b2b12301e0767d0d62e790534ca6661f7532133f934e93c2930fb"} Oct 01 16:02:37 crc kubenswrapper[4726]: E1001 16:02:37.049965 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 01 16:02:37 crc kubenswrapper[4726]: E1001 16:02:37.050657 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ncf6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-dk98l_openshift-marketplace(9a7a60ae-9178-467a-b52e-8e9c37dc64a3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 16:02:37 crc kubenswrapper[4726]: E1001 16:02:37.051862 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-dk98l" podUID="9a7a60ae-9178-467a-b52e-8e9c37dc64a3" Oct 01 16:02:37 crc kubenswrapper[4726]: E1001 16:02:37.075917 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 01 16:02:37 crc kubenswrapper[4726]: E1001 16:02:37.076168 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zzlxp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-nw2vt_openshift-marketplace(7502490c-e673-49e8-9645-775bdf4899df): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 16:02:37 crc kubenswrapper[4726]: E1001 16:02:37.077785 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-nw2vt" podUID="7502490c-e673-49e8-9645-775bdf4899df" Oct 01 16:02:37 crc kubenswrapper[4726]: I1001 16:02:37.328044 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gp78p" event={"ID":"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb","Type":"ContainerStarted","Data":"f4384c47fda3e134922a910a35b87db358ab77fe228643e65a0b4bb72496f25f"} Oct 01 16:02:37 crc kubenswrapper[4726]: I1001 16:02:37.331098 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" event={"ID":"2a22e7a8-726a-4538-a27d-44c58f89f0b1","Type":"ContainerStarted","Data":"968ebf7e7d8dd92ff12e58905cc807195d4eca5ae38ec1d7017d5620249659a3"} Oct 01 16:02:37 crc kubenswrapper[4726]: I1001 16:02:37.333650 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c","Type":"ContainerStarted","Data":"06cef0cf6b8dba81ef33cef4cb8cf75edfd8997b22746e913ebe90151a5ed11e"} Oct 01 16:02:37 crc kubenswrapper[4726]: E1001 16:02:37.338259 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-dk98l" podUID="9a7a60ae-9178-467a-b52e-8e9c37dc64a3" Oct 01 16:02:37 crc kubenswrapper[4726]: E1001 16:02:37.338509 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-nw2vt" podUID="7502490c-e673-49e8-9645-775bdf4899df" Oct 01 16:02:37 crc kubenswrapper[4726]: I1001 16:02:37.386044 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=48.386027668 podStartE2EDuration="48.386027668s" podCreationTimestamp="2025-10-01 16:01:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:02:37.370218982 +0000 UTC m=+210.271771569" watchObservedRunningTime="2025-10-01 16:02:37.386027668 +0000 UTC m=+210.287580245" Oct 01 16:02:38 crc kubenswrapper[4726]: I1001 16:02:38.341998 4726 generic.go:334] "Generic (PLEG): container finished" podID="21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c" containerID="06cef0cf6b8dba81ef33cef4cb8cf75edfd8997b22746e913ebe90151a5ed11e" exitCode=0 Oct 01 16:02:38 crc kubenswrapper[4726]: I1001 16:02:38.342103 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c","Type":"ContainerDied","Data":"06cef0cf6b8dba81ef33cef4cb8cf75edfd8997b22746e913ebe90151a5ed11e"} Oct 01 16:02:38 crc kubenswrapper[4726]: I1001 16:02:38.345511 4726 generic.go:334] "Generic (PLEG): container finished" podID="6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" containerID="f4384c47fda3e134922a910a35b87db358ab77fe228643e65a0b4bb72496f25f" exitCode=0 Oct 01 16:02:38 crc kubenswrapper[4726]: I1001 16:02:38.345586 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gp78p" event={"ID":"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb","Type":"ContainerDied","Data":"f4384c47fda3e134922a910a35b87db358ab77fe228643e65a0b4bb72496f25f"} Oct 01 16:02:38 crc kubenswrapper[4726]: I1001 16:02:38.347458 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4pzd8" event={"ID":"2a22e7a8-726a-4538-a27d-44c58f89f0b1","Type":"ContainerStarted","Data":"7bfb71683696c82adc3d99c4f1d7b234fcacb72168f81d91a8cecd05dfdd0c4b"} Oct 01 16:02:38 crc kubenswrapper[4726]: I1001 16:02:38.405284 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-4pzd8" podStartSLOduration=188.405264237 podStartE2EDuration="3m8.405264237s" podCreationTimestamp="2025-10-01 15:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:02:38.404649198 +0000 UTC m=+211.306201775" watchObservedRunningTime="2025-10-01 16:02:38.405264237 +0000 UTC m=+211.306816814" Oct 01 16:02:39 crc kubenswrapper[4726]: I1001 16:02:39.359446 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gp78p" event={"ID":"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb","Type":"ContainerStarted","Data":"39d0ecef3f802412950f66ef62210af69f6665b03c381ca927ee27977f71b9ce"} Oct 01 16:02:39 crc kubenswrapper[4726]: I1001 16:02:39.394535 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gp78p" podStartSLOduration=2.320589018 podStartE2EDuration="56.394505915s" podCreationTimestamp="2025-10-01 16:01:43 +0000 UTC" firstStartedPulling="2025-10-01 16:01:44.910753589 +0000 UTC m=+157.812306166" lastFinishedPulling="2025-10-01 16:02:38.984670486 +0000 UTC m=+211.886223063" observedRunningTime="2025-10-01 16:02:39.388598399 +0000 UTC m=+212.290151006" watchObservedRunningTime="2025-10-01 16:02:39.394505915 +0000 UTC m=+212.296058522" Oct 01 16:02:39 crc kubenswrapper[4726]: I1001 16:02:39.662998 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 16:02:39 crc kubenswrapper[4726]: I1001 16:02:39.732295 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c-kube-api-access\") pod \"21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c\" (UID: \"21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c\") " Oct 01 16:02:39 crc kubenswrapper[4726]: I1001 16:02:39.732397 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c-kubelet-dir\") pod \"21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c\" (UID: \"21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c\") " Oct 01 16:02:39 crc kubenswrapper[4726]: I1001 16:02:39.732561 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c" (UID: "21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:02:39 crc kubenswrapper[4726]: I1001 16:02:39.732786 4726 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:39 crc kubenswrapper[4726]: I1001 16:02:39.748423 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c" (UID: "21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:02:39 crc kubenswrapper[4726]: I1001 16:02:39.833974 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:40 crc kubenswrapper[4726]: I1001 16:02:40.364869 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 16:02:40 crc kubenswrapper[4726]: I1001 16:02:40.364855 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c","Type":"ContainerDied","Data":"a4e28a95a48db869d800c3f1c2a7f3dda48ecf4bc4abae6c63fe0d01b7ccb65a"} Oct 01 16:02:40 crc kubenswrapper[4726]: I1001 16:02:40.364922 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4e28a95a48db869d800c3f1c2a7f3dda48ecf4bc4abae6c63fe0d01b7ccb65a" Oct 01 16:02:42 crc kubenswrapper[4726]: I1001 16:02:42.375538 4726 generic.go:334] "Generic (PLEG): container finished" podID="03392899-81fb-49bf-8f23-8d056c3a8e19" containerID="e3013c589b20a10389c75239eb1db4e306b6c38abf99c34ec6e53c48b925d2e1" exitCode=0 Oct 01 16:02:42 crc kubenswrapper[4726]: I1001 16:02:42.375625 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qzjh" event={"ID":"03392899-81fb-49bf-8f23-8d056c3a8e19","Type":"ContainerDied","Data":"e3013c589b20a10389c75239eb1db4e306b6c38abf99c34ec6e53c48b925d2e1"} Oct 01 16:02:43 crc kubenswrapper[4726]: I1001 16:02:43.382831 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qzjh" event={"ID":"03392899-81fb-49bf-8f23-8d056c3a8e19","Type":"ContainerStarted","Data":"38462c42da536d937bb44beb1f5781029605444f080365737725be16dd5964a7"} Oct 01 16:02:43 crc kubenswrapper[4726]: I1001 16:02:43.402447 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4qzjh" podStartSLOduration=2.361103221 podStartE2EDuration="59.40242236s" podCreationTimestamp="2025-10-01 16:01:44 +0000 UTC" firstStartedPulling="2025-10-01 16:01:45.924538504 +0000 UTC m=+158.826091081" lastFinishedPulling="2025-10-01 16:02:42.965857633 +0000 UTC m=+215.867410220" observedRunningTime="2025-10-01 16:02:43.399179979 +0000 UTC m=+216.300732566" watchObservedRunningTime="2025-10-01 16:02:43.40242236 +0000 UTC m=+216.303974957" Oct 01 16:02:44 crc kubenswrapper[4726]: I1001 16:02:44.110773 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:02:44 crc kubenswrapper[4726]: I1001 16:02:44.111096 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:02:44 crc kubenswrapper[4726]: I1001 16:02:44.300059 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:02:44 crc kubenswrapper[4726]: I1001 16:02:44.390002 4726 generic.go:334] "Generic (PLEG): container finished" podID="9b53fe26-735b-47b1-b772-8de0898560ee" containerID="3f8f1ef1c46b122751f8053f8c04c9f938485487abda1aa8941ab2539127b195" exitCode=0 Oct 01 16:02:44 crc kubenswrapper[4726]: I1001 16:02:44.390754 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9m52k" event={"ID":"9b53fe26-735b-47b1-b772-8de0898560ee","Type":"ContainerDied","Data":"3f8f1ef1c46b122751f8053f8c04c9f938485487abda1aa8941ab2539127b195"} Oct 01 16:02:44 crc kubenswrapper[4726]: I1001 16:02:44.444394 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:02:44 crc kubenswrapper[4726]: I1001 16:02:44.527819 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:02:44 crc kubenswrapper[4726]: I1001 16:02:44.527892 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:02:45 crc kubenswrapper[4726]: I1001 16:02:45.399727 4726 generic.go:334] "Generic (PLEG): container finished" podID="8ecefa98-2ab6-4e4c-8a0b-589d845568f1" containerID="8c78bc1c419cc550dc84932322dab415790362e5dce614ba71c55195328f1361" exitCode=0 Oct 01 16:02:45 crc kubenswrapper[4726]: I1001 16:02:45.399827 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6v9s" event={"ID":"8ecefa98-2ab6-4e4c-8a0b-589d845568f1","Type":"ContainerDied","Data":"8c78bc1c419cc550dc84932322dab415790362e5dce614ba71c55195328f1361"} Oct 01 16:02:45 crc kubenswrapper[4726]: I1001 16:02:45.406579 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9m52k" event={"ID":"9b53fe26-735b-47b1-b772-8de0898560ee","Type":"ContainerStarted","Data":"588af8a2166cf90a46289e04b23a4e8abdc29d87734fff18b65f8d102656a08a"} Oct 01 16:02:45 crc kubenswrapper[4726]: I1001 16:02:45.445423 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9m52k" podStartSLOduration=2.562475579 podStartE2EDuration="1m0.445405317s" podCreationTimestamp="2025-10-01 16:01:45 +0000 UTC" firstStartedPulling="2025-10-01 16:01:46.962144084 +0000 UTC m=+159.863696661" lastFinishedPulling="2025-10-01 16:02:44.845073782 +0000 UTC m=+217.746626399" observedRunningTime="2025-10-01 16:02:45.441833925 +0000 UTC m=+218.343386512" watchObservedRunningTime="2025-10-01 16:02:45.445405317 +0000 UTC m=+218.346957884" Oct 01 16:02:45 crc kubenswrapper[4726]: I1001 16:02:45.572066 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-4qzjh" podUID="03392899-81fb-49bf-8f23-8d056c3a8e19" containerName="registry-server" probeResult="failure" output=< Oct 01 16:02:45 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Oct 01 16:02:45 crc kubenswrapper[4726]: > Oct 01 16:02:45 crc kubenswrapper[4726]: I1001 16:02:45.940434 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:02:45 crc kubenswrapper[4726]: I1001 16:02:45.940506 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:02:46 crc kubenswrapper[4726]: I1001 16:02:46.002491 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:02:46 crc kubenswrapper[4726]: I1001 16:02:46.416737 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6v9s" event={"ID":"8ecefa98-2ab6-4e4c-8a0b-589d845568f1","Type":"ContainerStarted","Data":"84f3f8ef74e10a1e6bee7f4a5bab5ab4d084356b6d5fe1c9f287d5d85dcba53a"} Oct 01 16:02:46 crc kubenswrapper[4726]: I1001 16:02:46.418736 4726 generic.go:334] "Generic (PLEG): container finished" podID="6406156d-358e-4c24-9e26-921d117dc024" containerID="6c25c118dc218e081d4ecb0749c39420b1ab543ac3a54a839dc7e28226af14cd" exitCode=0 Oct 01 16:02:46 crc kubenswrapper[4726]: I1001 16:02:46.418809 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hhbb4" event={"ID":"6406156d-358e-4c24-9e26-921d117dc024","Type":"ContainerDied","Data":"6c25c118dc218e081d4ecb0749c39420b1ab543ac3a54a839dc7e28226af14cd"} Oct 01 16:02:46 crc kubenswrapper[4726]: I1001 16:02:46.442528 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g6v9s" podStartSLOduration=2.485399316 podStartE2EDuration="1m1.442503711s" podCreationTimestamp="2025-10-01 16:01:45 +0000 UTC" firstStartedPulling="2025-10-01 16:01:46.956860253 +0000 UTC m=+159.858412830" lastFinishedPulling="2025-10-01 16:02:45.913964638 +0000 UTC m=+218.815517225" observedRunningTime="2025-10-01 16:02:46.440102765 +0000 UTC m=+219.341655362" watchObservedRunningTime="2025-10-01 16:02:46.442503711 +0000 UTC m=+219.344056318" Oct 01 16:02:47 crc kubenswrapper[4726]: I1001 16:02:47.428991 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hhbb4" event={"ID":"6406156d-358e-4c24-9e26-921d117dc024","Type":"ContainerStarted","Data":"b5dd4bff78f94c2869b76422340c3a24fae303e59d0a94380cb71ba4c7ea76d1"} Oct 01 16:02:47 crc kubenswrapper[4726]: I1001 16:02:47.465910 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hhbb4" podStartSLOduration=2.505858361 podStartE2EDuration="1m4.465881389s" podCreationTimestamp="2025-10-01 16:01:43 +0000 UTC" firstStartedPulling="2025-10-01 16:01:44.904372189 +0000 UTC m=+157.805924766" lastFinishedPulling="2025-10-01 16:02:46.864395207 +0000 UTC m=+219.765947794" observedRunningTime="2025-10-01 16:02:47.459942742 +0000 UTC m=+220.361495379" watchObservedRunningTime="2025-10-01 16:02:47.465881389 +0000 UTC m=+220.367434006" Oct 01 16:02:48 crc kubenswrapper[4726]: I1001 16:02:48.439714 4726 generic.go:334] "Generic (PLEG): container finished" podID="30b6165c-ee06-406a-b01e-35817eab4a32" containerID="2767bef2ca1a397b5f3f2d0a5a24d8864bea5fd37f6d6416c960ab8137b65f4e" exitCode=0 Oct 01 16:02:48 crc kubenswrapper[4726]: I1001 16:02:48.439818 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rfm7" event={"ID":"30b6165c-ee06-406a-b01e-35817eab4a32","Type":"ContainerDied","Data":"2767bef2ca1a397b5f3f2d0a5a24d8864bea5fd37f6d6416c960ab8137b65f4e"} Oct 01 16:02:49 crc kubenswrapper[4726]: I1001 16:02:49.450213 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rfm7" event={"ID":"30b6165c-ee06-406a-b01e-35817eab4a32","Type":"ContainerStarted","Data":"50aa7b23e8a995ef30763f4175fc48408121557a8fbdcdf3fa0ea4d0aa075670"} Oct 01 16:02:49 crc kubenswrapper[4726]: I1001 16:02:49.467788 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8rfm7" podStartSLOduration=3.530983141 podStartE2EDuration="1m6.467769306s" podCreationTimestamp="2025-10-01 16:01:43 +0000 UTC" firstStartedPulling="2025-10-01 16:01:45.920423335 +0000 UTC m=+158.821975912" lastFinishedPulling="2025-10-01 16:02:48.8572095 +0000 UTC m=+221.758762077" observedRunningTime="2025-10-01 16:02:49.46599497 +0000 UTC m=+222.367547567" watchObservedRunningTime="2025-10-01 16:02:49.467769306 +0000 UTC m=+222.369321893" Oct 01 16:02:52 crc kubenswrapper[4726]: I1001 16:02:52.473378 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dk98l" event={"ID":"9a7a60ae-9178-467a-b52e-8e9c37dc64a3","Type":"ContainerStarted","Data":"36d40cc9d96a1451565374389d7888507abb608901063d38197751ff3b8dc079"} Oct 01 16:02:52 crc kubenswrapper[4726]: I1001 16:02:52.479502 4726 generic.go:334] "Generic (PLEG): container finished" podID="7502490c-e673-49e8-9645-775bdf4899df" containerID="70c53cd74938333c951abb419be5f07c2f80589cc4867755a651b2a2db785ce4" exitCode=0 Oct 01 16:02:52 crc kubenswrapper[4726]: I1001 16:02:52.479568 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nw2vt" event={"ID":"7502490c-e673-49e8-9645-775bdf4899df","Type":"ContainerDied","Data":"70c53cd74938333c951abb419be5f07c2f80589cc4867755a651b2a2db785ce4"} Oct 01 16:02:53 crc kubenswrapper[4726]: I1001 16:02:53.413155 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:02:53 crc kubenswrapper[4726]: I1001 16:02:53.413538 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:02:53 crc kubenswrapper[4726]: I1001 16:02:53.413606 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 16:02:53 crc kubenswrapper[4726]: I1001 16:02:53.414336 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:02:53 crc kubenswrapper[4726]: I1001 16:02:53.414511 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6" gracePeriod=600 Oct 01 16:02:53 crc kubenswrapper[4726]: I1001 16:02:53.486840 4726 generic.go:334] "Generic (PLEG): container finished" podID="9a7a60ae-9178-467a-b52e-8e9c37dc64a3" containerID="36d40cc9d96a1451565374389d7888507abb608901063d38197751ff3b8dc079" exitCode=0 Oct 01 16:02:53 crc kubenswrapper[4726]: I1001 16:02:53.486891 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dk98l" event={"ID":"9a7a60ae-9178-467a-b52e-8e9c37dc64a3","Type":"ContainerDied","Data":"36d40cc9d96a1451565374389d7888507abb608901063d38197751ff3b8dc079"} Oct 01 16:02:53 crc kubenswrapper[4726]: I1001 16:02:53.978394 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:02:53 crc kubenswrapper[4726]: I1001 16:02:53.978464 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:02:54 crc kubenswrapper[4726]: I1001 16:02:54.033215 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:02:54 crc kubenswrapper[4726]: I1001 16:02:54.311892 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:02:54 crc kubenswrapper[4726]: I1001 16:02:54.312344 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:02:54 crc kubenswrapper[4726]: I1001 16:02:54.382458 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:02:54 crc kubenswrapper[4726]: I1001 16:02:54.494201 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6" exitCode=0 Oct 01 16:02:54 crc kubenswrapper[4726]: I1001 16:02:54.494312 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6"} Oct 01 16:02:54 crc kubenswrapper[4726]: I1001 16:02:54.549267 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:02:54 crc kubenswrapper[4726]: I1001 16:02:54.570807 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:02:54 crc kubenswrapper[4726]: I1001 16:02:54.614116 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:02:54 crc kubenswrapper[4726]: I1001 16:02:54.667381 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:02:55 crc kubenswrapper[4726]: I1001 16:02:55.989591 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:02:56 crc kubenswrapper[4726]: I1001 16:02:56.331302 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:02:56 crc kubenswrapper[4726]: I1001 16:02:56.331373 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:02:56 crc kubenswrapper[4726]: I1001 16:02:56.360584 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8rfm7"] Oct 01 16:02:56 crc kubenswrapper[4726]: I1001 16:02:56.385279 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:02:56 crc kubenswrapper[4726]: I1001 16:02:56.513171 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"29994399a03c1d7203fa51b4be7f3ba156ce8c41237e36d940e730558f1a7456"} Oct 01 16:02:56 crc kubenswrapper[4726]: I1001 16:02:56.513762 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8rfm7" podUID="30b6165c-ee06-406a-b01e-35817eab4a32" containerName="registry-server" containerID="cri-o://50aa7b23e8a995ef30763f4175fc48408121557a8fbdcdf3fa0ea4d0aa075670" gracePeriod=2 Oct 01 16:02:56 crc kubenswrapper[4726]: I1001 16:02:56.581533 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:02:57 crc kubenswrapper[4726]: I1001 16:02:57.519598 4726 generic.go:334] "Generic (PLEG): container finished" podID="30b6165c-ee06-406a-b01e-35817eab4a32" containerID="50aa7b23e8a995ef30763f4175fc48408121557a8fbdcdf3fa0ea4d0aa075670" exitCode=0 Oct 01 16:02:57 crc kubenswrapper[4726]: I1001 16:02:57.519707 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rfm7" event={"ID":"30b6165c-ee06-406a-b01e-35817eab4a32","Type":"ContainerDied","Data":"50aa7b23e8a995ef30763f4175fc48408121557a8fbdcdf3fa0ea4d0aa075670"} Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.100409 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.173862 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbqsm\" (UniqueName: \"kubernetes.io/projected/30b6165c-ee06-406a-b01e-35817eab4a32-kube-api-access-zbqsm\") pod \"30b6165c-ee06-406a-b01e-35817eab4a32\" (UID: \"30b6165c-ee06-406a-b01e-35817eab4a32\") " Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.173905 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b6165c-ee06-406a-b01e-35817eab4a32-catalog-content\") pod \"30b6165c-ee06-406a-b01e-35817eab4a32\" (UID: \"30b6165c-ee06-406a-b01e-35817eab4a32\") " Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.173990 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b6165c-ee06-406a-b01e-35817eab4a32-utilities\") pod \"30b6165c-ee06-406a-b01e-35817eab4a32\" (UID: \"30b6165c-ee06-406a-b01e-35817eab4a32\") " Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.175452 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30b6165c-ee06-406a-b01e-35817eab4a32-utilities" (OuterVolumeSpecName: "utilities") pod "30b6165c-ee06-406a-b01e-35817eab4a32" (UID: "30b6165c-ee06-406a-b01e-35817eab4a32"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.180830 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30b6165c-ee06-406a-b01e-35817eab4a32-kube-api-access-zbqsm" (OuterVolumeSpecName: "kube-api-access-zbqsm") pod "30b6165c-ee06-406a-b01e-35817eab4a32" (UID: "30b6165c-ee06-406a-b01e-35817eab4a32"). InnerVolumeSpecName "kube-api-access-zbqsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.217414 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30b6165c-ee06-406a-b01e-35817eab4a32-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30b6165c-ee06-406a-b01e-35817eab4a32" (UID: "30b6165c-ee06-406a-b01e-35817eab4a32"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.275641 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b6165c-ee06-406a-b01e-35817eab4a32-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.275678 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbqsm\" (UniqueName: \"kubernetes.io/projected/30b6165c-ee06-406a-b01e-35817eab4a32-kube-api-access-zbqsm\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.275689 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b6165c-ee06-406a-b01e-35817eab4a32-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.531262 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nw2vt" event={"ID":"7502490c-e673-49e8-9645-775bdf4899df","Type":"ContainerStarted","Data":"35889294da4c83d5ec528831cfa93a05430e8371a35c2c390e749f6b537f3d43"} Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.535708 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dk98l" event={"ID":"9a7a60ae-9178-467a-b52e-8e9c37dc64a3","Type":"ContainerStarted","Data":"9eea3a9c7c3f01e9f507b863a0763605eea5993b9a38e261c9b93eda402923b9"} Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.542566 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rfm7" event={"ID":"30b6165c-ee06-406a-b01e-35817eab4a32","Type":"ContainerDied","Data":"92b214e929c44e968c8e4f124511c320fc1e5cb086957e4f50d71d603f914508"} Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.542630 4726 scope.go:117] "RemoveContainer" containerID="50aa7b23e8a995ef30763f4175fc48408121557a8fbdcdf3fa0ea4d0aa075670" Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.542661 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rfm7" Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.559507 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nw2vt" podStartSLOduration=3.081993584 podStartE2EDuration="1m11.559487583s" podCreationTimestamp="2025-10-01 16:01:47 +0000 UTC" firstStartedPulling="2025-10-01 16:01:49.000161732 +0000 UTC m=+161.901714319" lastFinishedPulling="2025-10-01 16:02:57.477655711 +0000 UTC m=+230.379208318" observedRunningTime="2025-10-01 16:02:58.556436347 +0000 UTC m=+231.457988924" watchObservedRunningTime="2025-10-01 16:02:58.559487583 +0000 UTC m=+231.461040180" Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.565362 4726 scope.go:117] "RemoveContainer" containerID="2767bef2ca1a397b5f3f2d0a5a24d8864bea5fd37f6d6416c960ab8137b65f4e" Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.594711 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dk98l" podStartSLOduration=2.695322737 podStartE2EDuration="1m12.594690007s" podCreationTimestamp="2025-10-01 16:01:46 +0000 UTC" firstStartedPulling="2025-10-01 16:01:47.980359896 +0000 UTC m=+160.881912473" lastFinishedPulling="2025-10-01 16:02:57.879727166 +0000 UTC m=+230.781279743" observedRunningTime="2025-10-01 16:02:58.590369582 +0000 UTC m=+231.491922199" watchObservedRunningTime="2025-10-01 16:02:58.594690007 +0000 UTC m=+231.496242584" Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.602412 4726 scope.go:117] "RemoveContainer" containerID="7d6128006e57892a941cfc214d3dc9d0db634067bddcca8cff51d3f3fa1db45c" Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.612670 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8rfm7"] Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.619590 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8rfm7"] Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.756399 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4qzjh"] Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.756663 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4qzjh" podUID="03392899-81fb-49bf-8f23-8d056c3a8e19" containerName="registry-server" containerID="cri-o://38462c42da536d937bb44beb1f5781029605444f080365737725be16dd5964a7" gracePeriod=2 Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.957275 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6v9s"] Oct 01 16:02:58 crc kubenswrapper[4726]: I1001 16:02:58.957820 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g6v9s" podUID="8ecefa98-2ab6-4e4c-8a0b-589d845568f1" containerName="registry-server" containerID="cri-o://84f3f8ef74e10a1e6bee7f4a5bab5ab4d084356b6d5fe1c9f287d5d85dcba53a" gracePeriod=2 Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.105987 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.185124 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzxq8\" (UniqueName: \"kubernetes.io/projected/03392899-81fb-49bf-8f23-8d056c3a8e19-kube-api-access-rzxq8\") pod \"03392899-81fb-49bf-8f23-8d056c3a8e19\" (UID: \"03392899-81fb-49bf-8f23-8d056c3a8e19\") " Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.185529 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03392899-81fb-49bf-8f23-8d056c3a8e19-utilities\") pod \"03392899-81fb-49bf-8f23-8d056c3a8e19\" (UID: \"03392899-81fb-49bf-8f23-8d056c3a8e19\") " Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.185572 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03392899-81fb-49bf-8f23-8d056c3a8e19-catalog-content\") pod \"03392899-81fb-49bf-8f23-8d056c3a8e19\" (UID: \"03392899-81fb-49bf-8f23-8d056c3a8e19\") " Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.190652 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03392899-81fb-49bf-8f23-8d056c3a8e19-utilities" (OuterVolumeSpecName: "utilities") pod "03392899-81fb-49bf-8f23-8d056c3a8e19" (UID: "03392899-81fb-49bf-8f23-8d056c3a8e19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.193234 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03392899-81fb-49bf-8f23-8d056c3a8e19-kube-api-access-rzxq8" (OuterVolumeSpecName: "kube-api-access-rzxq8") pod "03392899-81fb-49bf-8f23-8d056c3a8e19" (UID: "03392899-81fb-49bf-8f23-8d056c3a8e19"). InnerVolumeSpecName "kube-api-access-rzxq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.286601 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03392899-81fb-49bf-8f23-8d056c3a8e19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03392899-81fb-49bf-8f23-8d056c3a8e19" (UID: "03392899-81fb-49bf-8f23-8d056c3a8e19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.287292 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03392899-81fb-49bf-8f23-8d056c3a8e19-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.287322 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzxq8\" (UniqueName: \"kubernetes.io/projected/03392899-81fb-49bf-8f23-8d056c3a8e19-kube-api-access-rzxq8\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.287333 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03392899-81fb-49bf-8f23-8d056c3a8e19-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.332300 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.388637 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-utilities\") pod \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\" (UID: \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\") " Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.388702 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-catalog-content\") pod \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\" (UID: \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\") " Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.388801 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcddw\" (UniqueName: \"kubernetes.io/projected/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-kube-api-access-vcddw\") pod \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\" (UID: \"8ecefa98-2ab6-4e4c-8a0b-589d845568f1\") " Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.389474 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-utilities" (OuterVolumeSpecName: "utilities") pod "8ecefa98-2ab6-4e4c-8a0b-589d845568f1" (UID: "8ecefa98-2ab6-4e4c-8a0b-589d845568f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.392282 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-kube-api-access-vcddw" (OuterVolumeSpecName: "kube-api-access-vcddw") pod "8ecefa98-2ab6-4e4c-8a0b-589d845568f1" (UID: "8ecefa98-2ab6-4e4c-8a0b-589d845568f1"). InnerVolumeSpecName "kube-api-access-vcddw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.412279 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ecefa98-2ab6-4e4c-8a0b-589d845568f1" (UID: "8ecefa98-2ab6-4e4c-8a0b-589d845568f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.490147 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcddw\" (UniqueName: \"kubernetes.io/projected/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-kube-api-access-vcddw\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.490383 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.490393 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ecefa98-2ab6-4e4c-8a0b-589d845568f1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.549524 4726 generic.go:334] "Generic (PLEG): container finished" podID="8ecefa98-2ab6-4e4c-8a0b-589d845568f1" containerID="84f3f8ef74e10a1e6bee7f4a5bab5ab4d084356b6d5fe1c9f287d5d85dcba53a" exitCode=0 Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.549561 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6v9s" event={"ID":"8ecefa98-2ab6-4e4c-8a0b-589d845568f1","Type":"ContainerDied","Data":"84f3f8ef74e10a1e6bee7f4a5bab5ab4d084356b6d5fe1c9f287d5d85dcba53a"} Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.549590 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g6v9s" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.549641 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6v9s" event={"ID":"8ecefa98-2ab6-4e4c-8a0b-589d845568f1","Type":"ContainerDied","Data":"2d71ab888087b0e8e6a5772bdc8468d17974f5506d6ffffce8258fb9ea64f9f8"} Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.549665 4726 scope.go:117] "RemoveContainer" containerID="84f3f8ef74e10a1e6bee7f4a5bab5ab4d084356b6d5fe1c9f287d5d85dcba53a" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.552405 4726 generic.go:334] "Generic (PLEG): container finished" podID="03392899-81fb-49bf-8f23-8d056c3a8e19" containerID="38462c42da536d937bb44beb1f5781029605444f080365737725be16dd5964a7" exitCode=0 Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.552507 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4qzjh" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.553167 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qzjh" event={"ID":"03392899-81fb-49bf-8f23-8d056c3a8e19","Type":"ContainerDied","Data":"38462c42da536d937bb44beb1f5781029605444f080365737725be16dd5964a7"} Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.553205 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qzjh" event={"ID":"03392899-81fb-49bf-8f23-8d056c3a8e19","Type":"ContainerDied","Data":"a3818d3414ab65ad0a5331b4e58c7873a25ddd17b9bd1b00d2968244d52ecbef"} Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.567089 4726 scope.go:117] "RemoveContainer" containerID="8c78bc1c419cc550dc84932322dab415790362e5dce614ba71c55195328f1361" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.593065 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4qzjh"] Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.599923 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4qzjh"] Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.608333 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6v9s"] Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.615432 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6v9s"] Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.616635 4726 scope.go:117] "RemoveContainer" containerID="669f5bf0a72e8665d8dd1dfc868c8db666bfe3d5fd8230273793fb982e95333d" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.631189 4726 scope.go:117] "RemoveContainer" containerID="84f3f8ef74e10a1e6bee7f4a5bab5ab4d084356b6d5fe1c9f287d5d85dcba53a" Oct 01 16:02:59 crc kubenswrapper[4726]: E1001 16:02:59.631650 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84f3f8ef74e10a1e6bee7f4a5bab5ab4d084356b6d5fe1c9f287d5d85dcba53a\": container with ID starting with 84f3f8ef74e10a1e6bee7f4a5bab5ab4d084356b6d5fe1c9f287d5d85dcba53a not found: ID does not exist" containerID="84f3f8ef74e10a1e6bee7f4a5bab5ab4d084356b6d5fe1c9f287d5d85dcba53a" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.631693 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f3f8ef74e10a1e6bee7f4a5bab5ab4d084356b6d5fe1c9f287d5d85dcba53a"} err="failed to get container status \"84f3f8ef74e10a1e6bee7f4a5bab5ab4d084356b6d5fe1c9f287d5d85dcba53a\": rpc error: code = NotFound desc = could not find container \"84f3f8ef74e10a1e6bee7f4a5bab5ab4d084356b6d5fe1c9f287d5d85dcba53a\": container with ID starting with 84f3f8ef74e10a1e6bee7f4a5bab5ab4d084356b6d5fe1c9f287d5d85dcba53a not found: ID does not exist" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.631722 4726 scope.go:117] "RemoveContainer" containerID="8c78bc1c419cc550dc84932322dab415790362e5dce614ba71c55195328f1361" Oct 01 16:02:59 crc kubenswrapper[4726]: E1001 16:02:59.632079 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c78bc1c419cc550dc84932322dab415790362e5dce614ba71c55195328f1361\": container with ID starting with 8c78bc1c419cc550dc84932322dab415790362e5dce614ba71c55195328f1361 not found: ID does not exist" containerID="8c78bc1c419cc550dc84932322dab415790362e5dce614ba71c55195328f1361" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.632100 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c78bc1c419cc550dc84932322dab415790362e5dce614ba71c55195328f1361"} err="failed to get container status \"8c78bc1c419cc550dc84932322dab415790362e5dce614ba71c55195328f1361\": rpc error: code = NotFound desc = could not find container \"8c78bc1c419cc550dc84932322dab415790362e5dce614ba71c55195328f1361\": container with ID starting with 8c78bc1c419cc550dc84932322dab415790362e5dce614ba71c55195328f1361 not found: ID does not exist" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.632114 4726 scope.go:117] "RemoveContainer" containerID="669f5bf0a72e8665d8dd1dfc868c8db666bfe3d5fd8230273793fb982e95333d" Oct 01 16:02:59 crc kubenswrapper[4726]: E1001 16:02:59.632326 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"669f5bf0a72e8665d8dd1dfc868c8db666bfe3d5fd8230273793fb982e95333d\": container with ID starting with 669f5bf0a72e8665d8dd1dfc868c8db666bfe3d5fd8230273793fb982e95333d not found: ID does not exist" containerID="669f5bf0a72e8665d8dd1dfc868c8db666bfe3d5fd8230273793fb982e95333d" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.632356 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"669f5bf0a72e8665d8dd1dfc868c8db666bfe3d5fd8230273793fb982e95333d"} err="failed to get container status \"669f5bf0a72e8665d8dd1dfc868c8db666bfe3d5fd8230273793fb982e95333d\": rpc error: code = NotFound desc = could not find container \"669f5bf0a72e8665d8dd1dfc868c8db666bfe3d5fd8230273793fb982e95333d\": container with ID starting with 669f5bf0a72e8665d8dd1dfc868c8db666bfe3d5fd8230273793fb982e95333d not found: ID does not exist" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.632377 4726 scope.go:117] "RemoveContainer" containerID="38462c42da536d937bb44beb1f5781029605444f080365737725be16dd5964a7" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.648612 4726 scope.go:117] "RemoveContainer" containerID="e3013c589b20a10389c75239eb1db4e306b6c38abf99c34ec6e53c48b925d2e1" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.662244 4726 scope.go:117] "RemoveContainer" containerID="308568c843e938acde8fe2bcb6e0e659cba7fd1db0e2fe88c19d626fc2ee2e99" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.681910 4726 scope.go:117] "RemoveContainer" containerID="38462c42da536d937bb44beb1f5781029605444f080365737725be16dd5964a7" Oct 01 16:02:59 crc kubenswrapper[4726]: E1001 16:02:59.682366 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38462c42da536d937bb44beb1f5781029605444f080365737725be16dd5964a7\": container with ID starting with 38462c42da536d937bb44beb1f5781029605444f080365737725be16dd5964a7 not found: ID does not exist" containerID="38462c42da536d937bb44beb1f5781029605444f080365737725be16dd5964a7" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.682452 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38462c42da536d937bb44beb1f5781029605444f080365737725be16dd5964a7"} err="failed to get container status \"38462c42da536d937bb44beb1f5781029605444f080365737725be16dd5964a7\": rpc error: code = NotFound desc = could not find container \"38462c42da536d937bb44beb1f5781029605444f080365737725be16dd5964a7\": container with ID starting with 38462c42da536d937bb44beb1f5781029605444f080365737725be16dd5964a7 not found: ID does not exist" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.682543 4726 scope.go:117] "RemoveContainer" containerID="e3013c589b20a10389c75239eb1db4e306b6c38abf99c34ec6e53c48b925d2e1" Oct 01 16:02:59 crc kubenswrapper[4726]: E1001 16:02:59.682874 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3013c589b20a10389c75239eb1db4e306b6c38abf99c34ec6e53c48b925d2e1\": container with ID starting with e3013c589b20a10389c75239eb1db4e306b6c38abf99c34ec6e53c48b925d2e1 not found: ID does not exist" containerID="e3013c589b20a10389c75239eb1db4e306b6c38abf99c34ec6e53c48b925d2e1" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.682973 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3013c589b20a10389c75239eb1db4e306b6c38abf99c34ec6e53c48b925d2e1"} err="failed to get container status \"e3013c589b20a10389c75239eb1db4e306b6c38abf99c34ec6e53c48b925d2e1\": rpc error: code = NotFound desc = could not find container \"e3013c589b20a10389c75239eb1db4e306b6c38abf99c34ec6e53c48b925d2e1\": container with ID starting with e3013c589b20a10389c75239eb1db4e306b6c38abf99c34ec6e53c48b925d2e1 not found: ID does not exist" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.683038 4726 scope.go:117] "RemoveContainer" containerID="308568c843e938acde8fe2bcb6e0e659cba7fd1db0e2fe88c19d626fc2ee2e99" Oct 01 16:02:59 crc kubenswrapper[4726]: E1001 16:02:59.683554 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"308568c843e938acde8fe2bcb6e0e659cba7fd1db0e2fe88c19d626fc2ee2e99\": container with ID starting with 308568c843e938acde8fe2bcb6e0e659cba7fd1db0e2fe88c19d626fc2ee2e99 not found: ID does not exist" containerID="308568c843e938acde8fe2bcb6e0e659cba7fd1db0e2fe88c19d626fc2ee2e99" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.683590 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"308568c843e938acde8fe2bcb6e0e659cba7fd1db0e2fe88c19d626fc2ee2e99"} err="failed to get container status \"308568c843e938acde8fe2bcb6e0e659cba7fd1db0e2fe88c19d626fc2ee2e99\": rpc error: code = NotFound desc = could not find container \"308568c843e938acde8fe2bcb6e0e659cba7fd1db0e2fe88c19d626fc2ee2e99\": container with ID starting with 308568c843e938acde8fe2bcb6e0e659cba7fd1db0e2fe88c19d626fc2ee2e99 not found: ID does not exist" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.814157 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03392899-81fb-49bf-8f23-8d056c3a8e19" path="/var/lib/kubelet/pods/03392899-81fb-49bf-8f23-8d056c3a8e19/volumes" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.814735 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30b6165c-ee06-406a-b01e-35817eab4a32" path="/var/lib/kubelet/pods/30b6165c-ee06-406a-b01e-35817eab4a32/volumes" Oct 01 16:02:59 crc kubenswrapper[4726]: I1001 16:02:59.815294 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ecefa98-2ab6-4e4c-8a0b-589d845568f1" path="/var/lib/kubelet/pods/8ecefa98-2ab6-4e4c-8a0b-589d845568f1/volumes" Oct 01 16:03:07 crc kubenswrapper[4726]: I1001 16:03:07.350190 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:03:07 crc kubenswrapper[4726]: I1001 16:03:07.350752 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:03:07 crc kubenswrapper[4726]: I1001 16:03:07.399876 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:03:07 crc kubenswrapper[4726]: I1001 16:03:07.637494 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:03:07 crc kubenswrapper[4726]: I1001 16:03:07.758114 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:03:07 crc kubenswrapper[4726]: I1001 16:03:07.758191 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:03:07 crc kubenswrapper[4726]: I1001 16:03:07.796742 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:03:08 crc kubenswrapper[4726]: I1001 16:03:08.662772 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:03:09 crc kubenswrapper[4726]: I1001 16:03:09.361503 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nw2vt"] Oct 01 16:03:10 crc kubenswrapper[4726]: I1001 16:03:10.612783 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nw2vt" podUID="7502490c-e673-49e8-9645-775bdf4899df" containerName="registry-server" containerID="cri-o://35889294da4c83d5ec528831cfa93a05430e8371a35c2c390e749f6b537f3d43" gracePeriod=2 Oct 01 16:03:11 crc kubenswrapper[4726]: I1001 16:03:11.619520 4726 generic.go:334] "Generic (PLEG): container finished" podID="7502490c-e673-49e8-9645-775bdf4899df" containerID="35889294da4c83d5ec528831cfa93a05430e8371a35c2c390e749f6b537f3d43" exitCode=0 Oct 01 16:03:11 crc kubenswrapper[4726]: I1001 16:03:11.619559 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nw2vt" event={"ID":"7502490c-e673-49e8-9645-775bdf4899df","Type":"ContainerDied","Data":"35889294da4c83d5ec528831cfa93a05430e8371a35c2c390e749f6b537f3d43"} Oct 01 16:03:11 crc kubenswrapper[4726]: I1001 16:03:11.864025 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:03:11 crc kubenswrapper[4726]: I1001 16:03:11.964335 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7502490c-e673-49e8-9645-775bdf4899df-utilities\") pod \"7502490c-e673-49e8-9645-775bdf4899df\" (UID: \"7502490c-e673-49e8-9645-775bdf4899df\") " Oct 01 16:03:11 crc kubenswrapper[4726]: I1001 16:03:11.964406 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7502490c-e673-49e8-9645-775bdf4899df-catalog-content\") pod \"7502490c-e673-49e8-9645-775bdf4899df\" (UID: \"7502490c-e673-49e8-9645-775bdf4899df\") " Oct 01 16:03:11 crc kubenswrapper[4726]: I1001 16:03:11.964497 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzlxp\" (UniqueName: \"kubernetes.io/projected/7502490c-e673-49e8-9645-775bdf4899df-kube-api-access-zzlxp\") pod \"7502490c-e673-49e8-9645-775bdf4899df\" (UID: \"7502490c-e673-49e8-9645-775bdf4899df\") " Oct 01 16:03:11 crc kubenswrapper[4726]: I1001 16:03:11.965735 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7502490c-e673-49e8-9645-775bdf4899df-utilities" (OuterVolumeSpecName: "utilities") pod "7502490c-e673-49e8-9645-775bdf4899df" (UID: "7502490c-e673-49e8-9645-775bdf4899df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:03:11 crc kubenswrapper[4726]: I1001 16:03:11.970318 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7502490c-e673-49e8-9645-775bdf4899df-kube-api-access-zzlxp" (OuterVolumeSpecName: "kube-api-access-zzlxp") pod "7502490c-e673-49e8-9645-775bdf4899df" (UID: "7502490c-e673-49e8-9645-775bdf4899df"). InnerVolumeSpecName "kube-api-access-zzlxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:12 crc kubenswrapper[4726]: I1001 16:03:12.042915 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7502490c-e673-49e8-9645-775bdf4899df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7502490c-e673-49e8-9645-775bdf4899df" (UID: "7502490c-e673-49e8-9645-775bdf4899df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:03:12 crc kubenswrapper[4726]: I1001 16:03:12.065680 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzlxp\" (UniqueName: \"kubernetes.io/projected/7502490c-e673-49e8-9645-775bdf4899df-kube-api-access-zzlxp\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:12 crc kubenswrapper[4726]: I1001 16:03:12.065722 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7502490c-e673-49e8-9645-775bdf4899df-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:12 crc kubenswrapper[4726]: I1001 16:03:12.065735 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7502490c-e673-49e8-9645-775bdf4899df-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:12 crc kubenswrapper[4726]: I1001 16:03:12.626095 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nw2vt" event={"ID":"7502490c-e673-49e8-9645-775bdf4899df","Type":"ContainerDied","Data":"5fd9a2bc5ee54c59f702c4f97cc87fce38f09969cc016fc5edb26d68f2c5e75a"} Oct 01 16:03:12 crc kubenswrapper[4726]: I1001 16:03:12.626152 4726 scope.go:117] "RemoveContainer" containerID="35889294da4c83d5ec528831cfa93a05430e8371a35c2c390e749f6b537f3d43" Oct 01 16:03:12 crc kubenswrapper[4726]: I1001 16:03:12.626292 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nw2vt" Oct 01 16:03:12 crc kubenswrapper[4726]: I1001 16:03:12.643143 4726 scope.go:117] "RemoveContainer" containerID="70c53cd74938333c951abb419be5f07c2f80589cc4867755a651b2a2db785ce4" Oct 01 16:03:12 crc kubenswrapper[4726]: I1001 16:03:12.654426 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nw2vt"] Oct 01 16:03:12 crc kubenswrapper[4726]: I1001 16:03:12.657300 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nw2vt"] Oct 01 16:03:12 crc kubenswrapper[4726]: I1001 16:03:12.659455 4726 scope.go:117] "RemoveContainer" containerID="8d927f5093b4a80920687a0d4b93b79daff81ee2e7255a6a55e764d380da56f7" Oct 01 16:03:13 crc kubenswrapper[4726]: I1001 16:03:13.814470 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7502490c-e673-49e8-9645-775bdf4899df" path="/var/lib/kubelet/pods/7502490c-e673-49e8-9645-775bdf4899df/volumes" Oct 01 16:03:16 crc kubenswrapper[4726]: I1001 16:03:16.619845 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pq5pm"] Oct 01 16:03:41 crc kubenswrapper[4726]: I1001 16:03:41.662172 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" podUID="875b60c0-2704-4bb1-9b01-5e6b0c745fa1" containerName="oauth-openshift" containerID="cri-o://79a70b9e2928f8d514e3032051c54621ab0bf490f16f6c0610ff45a767b6b90d" gracePeriod=15 Oct 01 16:03:41 crc kubenswrapper[4726]: I1001 16:03:41.832857 4726 generic.go:334] "Generic (PLEG): container finished" podID="875b60c0-2704-4bb1-9b01-5e6b0c745fa1" containerID="79a70b9e2928f8d514e3032051c54621ab0bf490f16f6c0610ff45a767b6b90d" exitCode=0 Oct 01 16:03:41 crc kubenswrapper[4726]: I1001 16:03:41.833468 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" event={"ID":"875b60c0-2704-4bb1-9b01-5e6b0c745fa1","Type":"ContainerDied","Data":"79a70b9e2928f8d514e3032051c54621ab0bf490f16f6c0610ff45a767b6b90d"} Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.068189 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101023 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-594b4c596d-vgmrp"] Oct 01 16:03:42 crc kubenswrapper[4726]: E1001 16:03:42.101322 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875b60c0-2704-4bb1-9b01-5e6b0c745fa1" containerName="oauth-openshift" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101344 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="875b60c0-2704-4bb1-9b01-5e6b0c745fa1" containerName="oauth-openshift" Oct 01 16:03:42 crc kubenswrapper[4726]: E1001 16:03:42.101365 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03392899-81fb-49bf-8f23-8d056c3a8e19" containerName="extract-utilities" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101373 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="03392899-81fb-49bf-8f23-8d056c3a8e19" containerName="extract-utilities" Oct 01 16:03:42 crc kubenswrapper[4726]: E1001 16:03:42.101383 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ecefa98-2ab6-4e4c-8a0b-589d845568f1" containerName="registry-server" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101393 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ecefa98-2ab6-4e4c-8a0b-589d845568f1" containerName="registry-server" Oct 01 16:03:42 crc kubenswrapper[4726]: E1001 16:03:42.101404 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7502490c-e673-49e8-9645-775bdf4899df" containerName="registry-server" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101413 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7502490c-e673-49e8-9645-775bdf4899df" containerName="registry-server" Oct 01 16:03:42 crc kubenswrapper[4726]: E1001 16:03:42.101424 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b6165c-ee06-406a-b01e-35817eab4a32" containerName="extract-utilities" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101432 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b6165c-ee06-406a-b01e-35817eab4a32" containerName="extract-utilities" Oct 01 16:03:42 crc kubenswrapper[4726]: E1001 16:03:42.101441 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03392899-81fb-49bf-8f23-8d056c3a8e19" containerName="registry-server" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101449 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="03392899-81fb-49bf-8f23-8d056c3a8e19" containerName="registry-server" Oct 01 16:03:42 crc kubenswrapper[4726]: E1001 16:03:42.101457 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b6165c-ee06-406a-b01e-35817eab4a32" containerName="extract-content" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101465 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b6165c-ee06-406a-b01e-35817eab4a32" containerName="extract-content" Oct 01 16:03:42 crc kubenswrapper[4726]: E1001 16:03:42.101477 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7502490c-e673-49e8-9645-775bdf4899df" containerName="extract-utilities" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101485 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7502490c-e673-49e8-9645-775bdf4899df" containerName="extract-utilities" Oct 01 16:03:42 crc kubenswrapper[4726]: E1001 16:03:42.101495 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7502490c-e673-49e8-9645-775bdf4899df" containerName="extract-content" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101504 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7502490c-e673-49e8-9645-775bdf4899df" containerName="extract-content" Oct 01 16:03:42 crc kubenswrapper[4726]: E1001 16:03:42.101515 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ecefa98-2ab6-4e4c-8a0b-589d845568f1" containerName="extract-utilities" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101523 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ecefa98-2ab6-4e4c-8a0b-589d845568f1" containerName="extract-utilities" Oct 01 16:03:42 crc kubenswrapper[4726]: E1001 16:03:42.101535 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b6165c-ee06-406a-b01e-35817eab4a32" containerName="registry-server" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101542 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b6165c-ee06-406a-b01e-35817eab4a32" containerName="registry-server" Oct 01 16:03:42 crc kubenswrapper[4726]: E1001 16:03:42.101555 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c" containerName="pruner" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101562 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c" containerName="pruner" Oct 01 16:03:42 crc kubenswrapper[4726]: E1001 16:03:42.101572 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ecefa98-2ab6-4e4c-8a0b-589d845568f1" containerName="extract-content" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101579 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ecefa98-2ab6-4e4c-8a0b-589d845568f1" containerName="extract-content" Oct 01 16:03:42 crc kubenswrapper[4726]: E1001 16:03:42.101591 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03392899-81fb-49bf-8f23-8d056c3a8e19" containerName="extract-content" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101599 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="03392899-81fb-49bf-8f23-8d056c3a8e19" containerName="extract-content" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101709 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="30b6165c-ee06-406a-b01e-35817eab4a32" containerName="registry-server" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101723 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="875b60c0-2704-4bb1-9b01-5e6b0c745fa1" containerName="oauth-openshift" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101737 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="7502490c-e673-49e8-9645-775bdf4899df" containerName="registry-server" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101747 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="21dc5d81-ff6f-4d8d-afc6-30d6dd87f15c" containerName="pruner" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101757 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ecefa98-2ab6-4e4c-8a0b-589d845568f1" containerName="registry-server" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.101769 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="03392899-81fb-49bf-8f23-8d056c3a8e19" containerName="registry-server" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.102366 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.109948 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-594b4c596d-vgmrp"] Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.162081 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-idp-0-file-data\") pod \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.162148 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-trusted-ca-bundle\") pod \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.162182 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-service-ca\") pod \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.162215 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-cliconfig\") pod \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.162243 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-audit-dir\") pod \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.162274 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6jp9\" (UniqueName: \"kubernetes.io/projected/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-kube-api-access-h6jp9\") pod \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.162306 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-router-certs\") pod \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.162352 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-serving-cert\") pod \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.162374 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-provider-selection\") pod \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.162408 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-session\") pod \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.162431 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-login\") pod \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.162486 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-ocp-branding-template\") pod \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.162519 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-error\") pod \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.162543 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-audit-policies\") pod \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\" (UID: \"875b60c0-2704-4bb1-9b01-5e6b0c745fa1\") " Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.163577 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "875b60c0-2704-4bb1-9b01-5e6b0c745fa1" (UID: "875b60c0-2704-4bb1-9b01-5e6b0c745fa1"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.164118 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "875b60c0-2704-4bb1-9b01-5e6b0c745fa1" (UID: "875b60c0-2704-4bb1-9b01-5e6b0c745fa1"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.164138 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "875b60c0-2704-4bb1-9b01-5e6b0c745fa1" (UID: "875b60c0-2704-4bb1-9b01-5e6b0c745fa1"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.164305 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "875b60c0-2704-4bb1-9b01-5e6b0c745fa1" (UID: "875b60c0-2704-4bb1-9b01-5e6b0c745fa1"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.164615 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "875b60c0-2704-4bb1-9b01-5e6b0c745fa1" (UID: "875b60c0-2704-4bb1-9b01-5e6b0c745fa1"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.170927 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "875b60c0-2704-4bb1-9b01-5e6b0c745fa1" (UID: "875b60c0-2704-4bb1-9b01-5e6b0c745fa1"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.171519 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "875b60c0-2704-4bb1-9b01-5e6b0c745fa1" (UID: "875b60c0-2704-4bb1-9b01-5e6b0c745fa1"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.172045 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "875b60c0-2704-4bb1-9b01-5e6b0c745fa1" (UID: "875b60c0-2704-4bb1-9b01-5e6b0c745fa1"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.172119 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-kube-api-access-h6jp9" (OuterVolumeSpecName: "kube-api-access-h6jp9") pod "875b60c0-2704-4bb1-9b01-5e6b0c745fa1" (UID: "875b60c0-2704-4bb1-9b01-5e6b0c745fa1"). InnerVolumeSpecName "kube-api-access-h6jp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.172641 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "875b60c0-2704-4bb1-9b01-5e6b0c745fa1" (UID: "875b60c0-2704-4bb1-9b01-5e6b0c745fa1"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.173035 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "875b60c0-2704-4bb1-9b01-5e6b0c745fa1" (UID: "875b60c0-2704-4bb1-9b01-5e6b0c745fa1"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.173225 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "875b60c0-2704-4bb1-9b01-5e6b0c745fa1" (UID: "875b60c0-2704-4bb1-9b01-5e6b0c745fa1"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.174410 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "875b60c0-2704-4bb1-9b01-5e6b0c745fa1" (UID: "875b60c0-2704-4bb1-9b01-5e6b0c745fa1"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.177468 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "875b60c0-2704-4bb1-9b01-5e6b0c745fa1" (UID: "875b60c0-2704-4bb1-9b01-5e6b0c745fa1"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.263406 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-serving-cert\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.263499 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.263538 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.263570 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-user-template-login\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.263658 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-cliconfig\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.263774 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.263825 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-session\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.263927 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fr8f\" (UniqueName: \"kubernetes.io/projected/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-kube-api-access-2fr8f\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.263988 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-audit-policies\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264007 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264030 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-service-ca\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264091 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-audit-dir\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264115 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-user-template-error\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264144 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-router-certs\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264224 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264239 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264251 4726 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264262 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264272 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264284 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264294 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264304 4726 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264314 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6jp9\" (UniqueName: \"kubernetes.io/projected/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-kube-api-access-h6jp9\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264323 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264332 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264341 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264352 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.264361 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/875b60c0-2704-4bb1-9b01-5e6b0c745fa1-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.364953 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fr8f\" (UniqueName: \"kubernetes.io/projected/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-kube-api-access-2fr8f\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.365023 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.365129 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-audit-policies\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.365159 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-service-ca\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.365186 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-audit-dir\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.365207 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-user-template-error\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.365234 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-router-certs\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.365260 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-serving-cert\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.365304 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.365330 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.365352 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-user-template-login\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.365378 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-cliconfig\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.365412 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.365435 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-session\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.365970 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-audit-dir\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.366677 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.367481 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-service-ca\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.367480 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-audit-policies\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.368147 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-cliconfig\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.369665 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-serving-cert\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.371079 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-user-template-error\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.371100 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-session\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.372117 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-user-template-login\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.372690 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-router-certs\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.373226 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.373604 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.373881 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.392095 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fr8f\" (UniqueName: \"kubernetes.io/projected/e90b26a3-a7e7-4c81-b63a-ec830a95ff55-kube-api-access-2fr8f\") pod \"oauth-openshift-594b4c596d-vgmrp\" (UID: \"e90b26a3-a7e7-4c81-b63a-ec830a95ff55\") " pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.435170 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.746979 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-594b4c596d-vgmrp"] Oct 01 16:03:42 crc kubenswrapper[4726]: W1001 16:03:42.752930 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode90b26a3_a7e7_4c81_b63a_ec830a95ff55.slice/crio-9e882f7ce3685d7d5fd25826e4c0e970c828a835ea2dff8192e70d03635335f3 WatchSource:0}: Error finding container 9e882f7ce3685d7d5fd25826e4c0e970c828a835ea2dff8192e70d03635335f3: Status 404 returned error can't find the container with id 9e882f7ce3685d7d5fd25826e4c0e970c828a835ea2dff8192e70d03635335f3 Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.844546 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" event={"ID":"875b60c0-2704-4bb1-9b01-5e6b0c745fa1","Type":"ContainerDied","Data":"b5937e733d4e1b1522e2779ff812f6aa8291624141c029958191cacab34d72cc"} Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.844637 4726 scope.go:117] "RemoveContainer" containerID="79a70b9e2928f8d514e3032051c54621ab0bf490f16f6c0610ff45a767b6b90d" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.844823 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pq5pm" Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.847454 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" event={"ID":"e90b26a3-a7e7-4c81-b63a-ec830a95ff55","Type":"ContainerStarted","Data":"9e882f7ce3685d7d5fd25826e4c0e970c828a835ea2dff8192e70d03635335f3"} Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.894989 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pq5pm"] Oct 01 16:03:42 crc kubenswrapper[4726]: I1001 16:03:42.899419 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pq5pm"] Oct 01 16:03:43 crc kubenswrapper[4726]: I1001 16:03:43.816606 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="875b60c0-2704-4bb1-9b01-5e6b0c745fa1" path="/var/lib/kubelet/pods/875b60c0-2704-4bb1-9b01-5e6b0c745fa1/volumes" Oct 01 16:03:43 crc kubenswrapper[4726]: I1001 16:03:43.853035 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" event={"ID":"e90b26a3-a7e7-4c81-b63a-ec830a95ff55","Type":"ContainerStarted","Data":"0f89ff0cb5c61242e3b81d210393c24d34a8c7c9313bdf7eae35a8b6c8fb7529"} Oct 01 16:03:43 crc kubenswrapper[4726]: I1001 16:03:43.853405 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:43 crc kubenswrapper[4726]: I1001 16:03:43.861187 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" Oct 01 16:03:43 crc kubenswrapper[4726]: I1001 16:03:43.871518 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-594b4c596d-vgmrp" podStartSLOduration=27.871500807 podStartE2EDuration="27.871500807s" podCreationTimestamp="2025-10-01 16:03:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:03:43.871020182 +0000 UTC m=+276.772572759" watchObservedRunningTime="2025-10-01 16:03:43.871500807 +0000 UTC m=+276.773053384" Oct 01 16:04:00 crc kubenswrapper[4726]: I1001 16:04:00.977843 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hhbb4"] Oct 01 16:04:00 crc kubenswrapper[4726]: I1001 16:04:00.978756 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hhbb4" podUID="6406156d-358e-4c24-9e26-921d117dc024" containerName="registry-server" containerID="cri-o://b5dd4bff78f94c2869b76422340c3a24fae303e59d0a94380cb71ba4c7ea76d1" gracePeriod=30 Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:00.999625 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gp78p"] Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:00.999911 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gp78p" podUID="6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" containerName="registry-server" containerID="cri-o://39d0ecef3f802412950f66ef62210af69f6665b03c381ca927ee27977f71b9ce" gracePeriod=30 Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.009192 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qvwb"] Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.009643 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" podUID="c442dae3-d66d-47a2-b1dc-30332e947203" containerName="marketplace-operator" containerID="cri-o://ac73ff1ef41d1fc551039f1f7f0e4b2e6e4585b60fcfc24370afcf7bd23c6eee" gracePeriod=30 Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.024073 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9m52k"] Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.024356 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9m52k" podUID="9b53fe26-735b-47b1-b772-8de0898560ee" containerName="registry-server" containerID="cri-o://588af8a2166cf90a46289e04b23a4e8abdc29d87734fff18b65f8d102656a08a" gracePeriod=30 Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.033154 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dk98l"] Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.033457 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dk98l" podUID="9a7a60ae-9178-467a-b52e-8e9c37dc64a3" containerName="registry-server" containerID="cri-o://9eea3a9c7c3f01e9f507b863a0763605eea5993b9a38e261c9b93eda402923b9" gracePeriod=30 Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.044714 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-g69q5"] Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.045490 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.045785 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-g69q5"] Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.233395 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c460d271-b44c-482b-a0d9-70bdc879cf1d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-g69q5\" (UID: \"c460d271-b44c-482b-a0d9-70bdc879cf1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.233741 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c460d271-b44c-482b-a0d9-70bdc879cf1d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-g69q5\" (UID: \"c460d271-b44c-482b-a0d9-70bdc879cf1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.233804 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pplm\" (UniqueName: \"kubernetes.io/projected/c460d271-b44c-482b-a0d9-70bdc879cf1d-kube-api-access-9pplm\") pod \"marketplace-operator-79b997595-g69q5\" (UID: \"c460d271-b44c-482b-a0d9-70bdc879cf1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.336466 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c460d271-b44c-482b-a0d9-70bdc879cf1d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-g69q5\" (UID: \"c460d271-b44c-482b-a0d9-70bdc879cf1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.336580 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c460d271-b44c-482b-a0d9-70bdc879cf1d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-g69q5\" (UID: \"c460d271-b44c-482b-a0d9-70bdc879cf1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.336670 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pplm\" (UniqueName: \"kubernetes.io/projected/c460d271-b44c-482b-a0d9-70bdc879cf1d-kube-api-access-9pplm\") pod \"marketplace-operator-79b997595-g69q5\" (UID: \"c460d271-b44c-482b-a0d9-70bdc879cf1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.340436 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c460d271-b44c-482b-a0d9-70bdc879cf1d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-g69q5\" (UID: \"c460d271-b44c-482b-a0d9-70bdc879cf1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.342982 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c460d271-b44c-482b-a0d9-70bdc879cf1d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-g69q5\" (UID: \"c460d271-b44c-482b-a0d9-70bdc879cf1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.359441 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pplm\" (UniqueName: \"kubernetes.io/projected/c460d271-b44c-482b-a0d9-70bdc879cf1d-kube-api-access-9pplm\") pod \"marketplace-operator-79b997595-g69q5\" (UID: \"c460d271-b44c-482b-a0d9-70bdc879cf1d\") " pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.444522 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.450091 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.452304 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.458243 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.461823 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.471865 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639286 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6406156d-358e-4c24-9e26-921d117dc024-catalog-content\") pod \"6406156d-358e-4c24-9e26-921d117dc024\" (UID: \"6406156d-358e-4c24-9e26-921d117dc024\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639554 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dx8qg\" (UniqueName: \"kubernetes.io/projected/c442dae3-d66d-47a2-b1dc-30332e947203-kube-api-access-dx8qg\") pod \"c442dae3-d66d-47a2-b1dc-30332e947203\" (UID: \"c442dae3-d66d-47a2-b1dc-30332e947203\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639592 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c442dae3-d66d-47a2-b1dc-30332e947203-marketplace-trusted-ca\") pod \"c442dae3-d66d-47a2-b1dc-30332e947203\" (UID: \"c442dae3-d66d-47a2-b1dc-30332e947203\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639622 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6406156d-358e-4c24-9e26-921d117dc024-utilities\") pod \"6406156d-358e-4c24-9e26-921d117dc024\" (UID: \"6406156d-358e-4c24-9e26-921d117dc024\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639666 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntz7h\" (UniqueName: \"kubernetes.io/projected/6406156d-358e-4c24-9e26-921d117dc024-kube-api-access-ntz7h\") pod \"6406156d-358e-4c24-9e26-921d117dc024\" (UID: \"6406156d-358e-4c24-9e26-921d117dc024\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639689 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-utilities\") pod \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\" (UID: \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639724 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncf6z\" (UniqueName: \"kubernetes.io/projected/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-kube-api-access-ncf6z\") pod \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\" (UID: \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639756 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b53fe26-735b-47b1-b772-8de0898560ee-catalog-content\") pod \"9b53fe26-735b-47b1-b772-8de0898560ee\" (UID: \"9b53fe26-735b-47b1-b772-8de0898560ee\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639780 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6b9s\" (UniqueName: \"kubernetes.io/projected/9b53fe26-735b-47b1-b772-8de0898560ee-kube-api-access-n6b9s\") pod \"9b53fe26-735b-47b1-b772-8de0898560ee\" (UID: \"9b53fe26-735b-47b1-b772-8de0898560ee\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639809 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c442dae3-d66d-47a2-b1dc-30332e947203-marketplace-operator-metrics\") pod \"c442dae3-d66d-47a2-b1dc-30332e947203\" (UID: \"c442dae3-d66d-47a2-b1dc-30332e947203\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639847 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-utilities\") pod \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\" (UID: \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639869 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b53fe26-735b-47b1-b772-8de0898560ee-utilities\") pod \"9b53fe26-735b-47b1-b772-8de0898560ee\" (UID: \"9b53fe26-735b-47b1-b772-8de0898560ee\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639892 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-catalog-content\") pod \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\" (UID: \"9a7a60ae-9178-467a-b52e-8e9c37dc64a3\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639915 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g482\" (UniqueName: \"kubernetes.io/projected/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-kube-api-access-6g482\") pod \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\" (UID: \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.639948 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-catalog-content\") pod \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\" (UID: \"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb\") " Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.640711 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6406156d-358e-4c24-9e26-921d117dc024-utilities" (OuterVolumeSpecName: "utilities") pod "6406156d-358e-4c24-9e26-921d117dc024" (UID: "6406156d-358e-4c24-9e26-921d117dc024"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.640958 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c442dae3-d66d-47a2-b1dc-30332e947203-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "c442dae3-d66d-47a2-b1dc-30332e947203" (UID: "c442dae3-d66d-47a2-b1dc-30332e947203"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.641744 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-utilities" (OuterVolumeSpecName: "utilities") pod "6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" (UID: "6dae1d83-2eb4-4f56-b14d-1d46fd6908eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.641979 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-utilities" (OuterVolumeSpecName: "utilities") pod "9a7a60ae-9178-467a-b52e-8e9c37dc64a3" (UID: "9a7a60ae-9178-467a-b52e-8e9c37dc64a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.641974 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b53fe26-735b-47b1-b772-8de0898560ee-utilities" (OuterVolumeSpecName: "utilities") pod "9b53fe26-735b-47b1-b772-8de0898560ee" (UID: "9b53fe26-735b-47b1-b772-8de0898560ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.645635 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b53fe26-735b-47b1-b772-8de0898560ee-kube-api-access-n6b9s" (OuterVolumeSpecName: "kube-api-access-n6b9s") pod "9b53fe26-735b-47b1-b772-8de0898560ee" (UID: "9b53fe26-735b-47b1-b772-8de0898560ee"). InnerVolumeSpecName "kube-api-access-n6b9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.650694 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6406156d-358e-4c24-9e26-921d117dc024-kube-api-access-ntz7h" (OuterVolumeSpecName: "kube-api-access-ntz7h") pod "6406156d-358e-4c24-9e26-921d117dc024" (UID: "6406156d-358e-4c24-9e26-921d117dc024"). InnerVolumeSpecName "kube-api-access-ntz7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.654721 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c442dae3-d66d-47a2-b1dc-30332e947203-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "c442dae3-d66d-47a2-b1dc-30332e947203" (UID: "c442dae3-d66d-47a2-b1dc-30332e947203"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.655661 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-kube-api-access-ncf6z" (OuterVolumeSpecName: "kube-api-access-ncf6z") pod "9a7a60ae-9178-467a-b52e-8e9c37dc64a3" (UID: "9a7a60ae-9178-467a-b52e-8e9c37dc64a3"). InnerVolumeSpecName "kube-api-access-ncf6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.657413 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c442dae3-d66d-47a2-b1dc-30332e947203-kube-api-access-dx8qg" (OuterVolumeSpecName: "kube-api-access-dx8qg") pod "c442dae3-d66d-47a2-b1dc-30332e947203" (UID: "c442dae3-d66d-47a2-b1dc-30332e947203"). InnerVolumeSpecName "kube-api-access-dx8qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.657734 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-kube-api-access-6g482" (OuterVolumeSpecName: "kube-api-access-6g482") pod "6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" (UID: "6dae1d83-2eb4-4f56-b14d-1d46fd6908eb"). InnerVolumeSpecName "kube-api-access-6g482". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.684358 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-g69q5"] Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.686570 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b53fe26-735b-47b1-b772-8de0898560ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9b53fe26-735b-47b1-b772-8de0898560ee" (UID: "9b53fe26-735b-47b1-b772-8de0898560ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.722185 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6406156d-358e-4c24-9e26-921d117dc024-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6406156d-358e-4c24-9e26-921d117dc024" (UID: "6406156d-358e-4c24-9e26-921d117dc024"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.734032 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" (UID: "6dae1d83-2eb4-4f56-b14d-1d46fd6908eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.743766 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b53fe26-735b-47b1-b772-8de0898560ee-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.743796 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g482\" (UniqueName: \"kubernetes.io/projected/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-kube-api-access-6g482\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.743807 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.743816 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6406156d-358e-4c24-9e26-921d117dc024-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.743825 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dx8qg\" (UniqueName: \"kubernetes.io/projected/c442dae3-d66d-47a2-b1dc-30332e947203-kube-api-access-dx8qg\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.743834 4726 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c442dae3-d66d-47a2-b1dc-30332e947203-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.743842 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6406156d-358e-4c24-9e26-921d117dc024-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.743850 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntz7h\" (UniqueName: \"kubernetes.io/projected/6406156d-358e-4c24-9e26-921d117dc024-kube-api-access-ntz7h\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.743858 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.743866 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncf6z\" (UniqueName: \"kubernetes.io/projected/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-kube-api-access-ncf6z\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.743874 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b53fe26-735b-47b1-b772-8de0898560ee-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.743881 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6b9s\" (UniqueName: \"kubernetes.io/projected/9b53fe26-735b-47b1-b772-8de0898560ee-kube-api-access-n6b9s\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.743889 4726 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c442dae3-d66d-47a2-b1dc-30332e947203-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.743898 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.749748 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a7a60ae-9178-467a-b52e-8e9c37dc64a3" (UID: "9a7a60ae-9178-467a-b52e-8e9c37dc64a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.845600 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7a60ae-9178-467a-b52e-8e9c37dc64a3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.964216 4726 generic.go:334] "Generic (PLEG): container finished" podID="9b53fe26-735b-47b1-b772-8de0898560ee" containerID="588af8a2166cf90a46289e04b23a4e8abdc29d87734fff18b65f8d102656a08a" exitCode=0 Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.964295 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9m52k" event={"ID":"9b53fe26-735b-47b1-b772-8de0898560ee","Type":"ContainerDied","Data":"588af8a2166cf90a46289e04b23a4e8abdc29d87734fff18b65f8d102656a08a"} Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.964317 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9m52k" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.964329 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9m52k" event={"ID":"9b53fe26-735b-47b1-b772-8de0898560ee","Type":"ContainerDied","Data":"cb22d203e077bb599e6b97969e2a3d5df12c11b62ae2f0adaa67d97c4a9252fa"} Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.964351 4726 scope.go:117] "RemoveContainer" containerID="588af8a2166cf90a46289e04b23a4e8abdc29d87734fff18b65f8d102656a08a" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.968123 4726 generic.go:334] "Generic (PLEG): container finished" podID="6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" containerID="39d0ecef3f802412950f66ef62210af69f6665b03c381ca927ee27977f71b9ce" exitCode=0 Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.968196 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gp78p" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.968202 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gp78p" event={"ID":"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb","Type":"ContainerDied","Data":"39d0ecef3f802412950f66ef62210af69f6665b03c381ca927ee27977f71b9ce"} Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.968252 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gp78p" event={"ID":"6dae1d83-2eb4-4f56-b14d-1d46fd6908eb","Type":"ContainerDied","Data":"571b5db544d7c381d35c4249c1dbe752165549ec94063bc9b7e86f23e1b0732f"} Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.971031 4726 generic.go:334] "Generic (PLEG): container finished" podID="c442dae3-d66d-47a2-b1dc-30332e947203" containerID="ac73ff1ef41d1fc551039f1f7f0e4b2e6e4585b60fcfc24370afcf7bd23c6eee" exitCode=0 Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.971110 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" event={"ID":"c442dae3-d66d-47a2-b1dc-30332e947203","Type":"ContainerDied","Data":"ac73ff1ef41d1fc551039f1f7f0e4b2e6e4585b60fcfc24370afcf7bd23c6eee"} Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.971127 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" event={"ID":"c442dae3-d66d-47a2-b1dc-30332e947203","Type":"ContainerDied","Data":"0e73fc792d3c86d3340a0208a3f4725c6f8e7a9dd4082e7049870f04bbed0876"} Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.971196 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9qvwb" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.974034 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" event={"ID":"c460d271-b44c-482b-a0d9-70bdc879cf1d","Type":"ContainerStarted","Data":"6b95d8af151bf60d230fd634a55919dfbf75aacf27da5fe1ad05cd42662cbc59"} Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.974111 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" event={"ID":"c460d271-b44c-482b-a0d9-70bdc879cf1d","Type":"ContainerStarted","Data":"f10ae322949fe43be064a1234b3efa0f82730bd1337d307182cf6a60c54b7b5a"} Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.974600 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.975646 4726 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-g69q5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.975687 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" podUID="c460d271-b44c-482b-a0d9-70bdc879cf1d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.978933 4726 generic.go:334] "Generic (PLEG): container finished" podID="6406156d-358e-4c24-9e26-921d117dc024" containerID="b5dd4bff78f94c2869b76422340c3a24fae303e59d0a94380cb71ba4c7ea76d1" exitCode=0 Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.979009 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hhbb4" event={"ID":"6406156d-358e-4c24-9e26-921d117dc024","Type":"ContainerDied","Data":"b5dd4bff78f94c2869b76422340c3a24fae303e59d0a94380cb71ba4c7ea76d1"} Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.979033 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hhbb4" event={"ID":"6406156d-358e-4c24-9e26-921d117dc024","Type":"ContainerDied","Data":"38098a5c023d48324aa8e5a4b494b2576d34934411107b77105c54166c920a14"} Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.979138 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hhbb4" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.982750 4726 generic.go:334] "Generic (PLEG): container finished" podID="9a7a60ae-9178-467a-b52e-8e9c37dc64a3" containerID="9eea3a9c7c3f01e9f507b863a0763605eea5993b9a38e261c9b93eda402923b9" exitCode=0 Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.982819 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dk98l" event={"ID":"9a7a60ae-9178-467a-b52e-8e9c37dc64a3","Type":"ContainerDied","Data":"9eea3a9c7c3f01e9f507b863a0763605eea5993b9a38e261c9b93eda402923b9"} Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.982844 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dk98l" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.982870 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dk98l" event={"ID":"9a7a60ae-9178-467a-b52e-8e9c37dc64a3","Type":"ContainerDied","Data":"b10e4c3898eb00bc4f3c6e51473710cdc784636f60a7b8e80d9e6838b03bbac2"} Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.985481 4726 scope.go:117] "RemoveContainer" containerID="3f8f1ef1c46b122751f8053f8c04c9f938485487abda1aa8941ab2539127b195" Oct 01 16:04:01 crc kubenswrapper[4726]: I1001 16:04:01.991778 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" podStartSLOduration=0.991746731 podStartE2EDuration="991.746731ms" podCreationTimestamp="2025-10-01 16:04:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:01.991110882 +0000 UTC m=+294.892663469" watchObservedRunningTime="2025-10-01 16:04:01.991746731 +0000 UTC m=+294.893299308" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.006519 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9m52k"] Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.013254 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9m52k"] Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.013392 4726 scope.go:117] "RemoveContainer" containerID="c477b1235eeba774a5ea24f37c4bfb8c59e5618f8784fadf2962617e71a3618c" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.024268 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hhbb4"] Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.027425 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hhbb4"] Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.038643 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gp78p"] Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.038767 4726 scope.go:117] "RemoveContainer" containerID="588af8a2166cf90a46289e04b23a4e8abdc29d87734fff18b65f8d102656a08a" Oct 01 16:04:02 crc kubenswrapper[4726]: E1001 16:04:02.039213 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"588af8a2166cf90a46289e04b23a4e8abdc29d87734fff18b65f8d102656a08a\": container with ID starting with 588af8a2166cf90a46289e04b23a4e8abdc29d87734fff18b65f8d102656a08a not found: ID does not exist" containerID="588af8a2166cf90a46289e04b23a4e8abdc29d87734fff18b65f8d102656a08a" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.039249 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"588af8a2166cf90a46289e04b23a4e8abdc29d87734fff18b65f8d102656a08a"} err="failed to get container status \"588af8a2166cf90a46289e04b23a4e8abdc29d87734fff18b65f8d102656a08a\": rpc error: code = NotFound desc = could not find container \"588af8a2166cf90a46289e04b23a4e8abdc29d87734fff18b65f8d102656a08a\": container with ID starting with 588af8a2166cf90a46289e04b23a4e8abdc29d87734fff18b65f8d102656a08a not found: ID does not exist" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.039279 4726 scope.go:117] "RemoveContainer" containerID="3f8f1ef1c46b122751f8053f8c04c9f938485487abda1aa8941ab2539127b195" Oct 01 16:04:02 crc kubenswrapper[4726]: E1001 16:04:02.039713 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f8f1ef1c46b122751f8053f8c04c9f938485487abda1aa8941ab2539127b195\": container with ID starting with 3f8f1ef1c46b122751f8053f8c04c9f938485487abda1aa8941ab2539127b195 not found: ID does not exist" containerID="3f8f1ef1c46b122751f8053f8c04c9f938485487abda1aa8941ab2539127b195" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.039768 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f8f1ef1c46b122751f8053f8c04c9f938485487abda1aa8941ab2539127b195"} err="failed to get container status \"3f8f1ef1c46b122751f8053f8c04c9f938485487abda1aa8941ab2539127b195\": rpc error: code = NotFound desc = could not find container \"3f8f1ef1c46b122751f8053f8c04c9f938485487abda1aa8941ab2539127b195\": container with ID starting with 3f8f1ef1c46b122751f8053f8c04c9f938485487abda1aa8941ab2539127b195 not found: ID does not exist" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.039796 4726 scope.go:117] "RemoveContainer" containerID="c477b1235eeba774a5ea24f37c4bfb8c59e5618f8784fadf2962617e71a3618c" Oct 01 16:04:02 crc kubenswrapper[4726]: E1001 16:04:02.040110 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c477b1235eeba774a5ea24f37c4bfb8c59e5618f8784fadf2962617e71a3618c\": container with ID starting with c477b1235eeba774a5ea24f37c4bfb8c59e5618f8784fadf2962617e71a3618c not found: ID does not exist" containerID="c477b1235eeba774a5ea24f37c4bfb8c59e5618f8784fadf2962617e71a3618c" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.040145 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c477b1235eeba774a5ea24f37c4bfb8c59e5618f8784fadf2962617e71a3618c"} err="failed to get container status \"c477b1235eeba774a5ea24f37c4bfb8c59e5618f8784fadf2962617e71a3618c\": rpc error: code = NotFound desc = could not find container \"c477b1235eeba774a5ea24f37c4bfb8c59e5618f8784fadf2962617e71a3618c\": container with ID starting with c477b1235eeba774a5ea24f37c4bfb8c59e5618f8784fadf2962617e71a3618c not found: ID does not exist" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.040179 4726 scope.go:117] "RemoveContainer" containerID="39d0ecef3f802412950f66ef62210af69f6665b03c381ca927ee27977f71b9ce" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.041377 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gp78p"] Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.046101 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qvwb"] Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.054311 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qvwb"] Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.055211 4726 scope.go:117] "RemoveContainer" containerID="f4384c47fda3e134922a910a35b87db358ab77fe228643e65a0b4bb72496f25f" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.062577 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dk98l"] Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.064613 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dk98l"] Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.067638 4726 scope.go:117] "RemoveContainer" containerID="8a8825e1ae9ee2b2363e5d02ad358f197b868264660b6da6403a18355b6cf3d3" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.079869 4726 scope.go:117] "RemoveContainer" containerID="39d0ecef3f802412950f66ef62210af69f6665b03c381ca927ee27977f71b9ce" Oct 01 16:04:02 crc kubenswrapper[4726]: E1001 16:04:02.080512 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39d0ecef3f802412950f66ef62210af69f6665b03c381ca927ee27977f71b9ce\": container with ID starting with 39d0ecef3f802412950f66ef62210af69f6665b03c381ca927ee27977f71b9ce not found: ID does not exist" containerID="39d0ecef3f802412950f66ef62210af69f6665b03c381ca927ee27977f71b9ce" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.080547 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39d0ecef3f802412950f66ef62210af69f6665b03c381ca927ee27977f71b9ce"} err="failed to get container status \"39d0ecef3f802412950f66ef62210af69f6665b03c381ca927ee27977f71b9ce\": rpc error: code = NotFound desc = could not find container \"39d0ecef3f802412950f66ef62210af69f6665b03c381ca927ee27977f71b9ce\": container with ID starting with 39d0ecef3f802412950f66ef62210af69f6665b03c381ca927ee27977f71b9ce not found: ID does not exist" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.080574 4726 scope.go:117] "RemoveContainer" containerID="f4384c47fda3e134922a910a35b87db358ab77fe228643e65a0b4bb72496f25f" Oct 01 16:04:02 crc kubenswrapper[4726]: E1001 16:04:02.080802 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4384c47fda3e134922a910a35b87db358ab77fe228643e65a0b4bb72496f25f\": container with ID starting with f4384c47fda3e134922a910a35b87db358ab77fe228643e65a0b4bb72496f25f not found: ID does not exist" containerID="f4384c47fda3e134922a910a35b87db358ab77fe228643e65a0b4bb72496f25f" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.080829 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4384c47fda3e134922a910a35b87db358ab77fe228643e65a0b4bb72496f25f"} err="failed to get container status \"f4384c47fda3e134922a910a35b87db358ab77fe228643e65a0b4bb72496f25f\": rpc error: code = NotFound desc = could not find container \"f4384c47fda3e134922a910a35b87db358ab77fe228643e65a0b4bb72496f25f\": container with ID starting with f4384c47fda3e134922a910a35b87db358ab77fe228643e65a0b4bb72496f25f not found: ID does not exist" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.080847 4726 scope.go:117] "RemoveContainer" containerID="8a8825e1ae9ee2b2363e5d02ad358f197b868264660b6da6403a18355b6cf3d3" Oct 01 16:04:02 crc kubenswrapper[4726]: E1001 16:04:02.081068 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a8825e1ae9ee2b2363e5d02ad358f197b868264660b6da6403a18355b6cf3d3\": container with ID starting with 8a8825e1ae9ee2b2363e5d02ad358f197b868264660b6da6403a18355b6cf3d3 not found: ID does not exist" containerID="8a8825e1ae9ee2b2363e5d02ad358f197b868264660b6da6403a18355b6cf3d3" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.081098 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a8825e1ae9ee2b2363e5d02ad358f197b868264660b6da6403a18355b6cf3d3"} err="failed to get container status \"8a8825e1ae9ee2b2363e5d02ad358f197b868264660b6da6403a18355b6cf3d3\": rpc error: code = NotFound desc = could not find container \"8a8825e1ae9ee2b2363e5d02ad358f197b868264660b6da6403a18355b6cf3d3\": container with ID starting with 8a8825e1ae9ee2b2363e5d02ad358f197b868264660b6da6403a18355b6cf3d3 not found: ID does not exist" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.081116 4726 scope.go:117] "RemoveContainer" containerID="ac73ff1ef41d1fc551039f1f7f0e4b2e6e4585b60fcfc24370afcf7bd23c6eee" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.091940 4726 scope.go:117] "RemoveContainer" containerID="ac73ff1ef41d1fc551039f1f7f0e4b2e6e4585b60fcfc24370afcf7bd23c6eee" Oct 01 16:04:02 crc kubenswrapper[4726]: E1001 16:04:02.092371 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac73ff1ef41d1fc551039f1f7f0e4b2e6e4585b60fcfc24370afcf7bd23c6eee\": container with ID starting with ac73ff1ef41d1fc551039f1f7f0e4b2e6e4585b60fcfc24370afcf7bd23c6eee not found: ID does not exist" containerID="ac73ff1ef41d1fc551039f1f7f0e4b2e6e4585b60fcfc24370afcf7bd23c6eee" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.092402 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac73ff1ef41d1fc551039f1f7f0e4b2e6e4585b60fcfc24370afcf7bd23c6eee"} err="failed to get container status \"ac73ff1ef41d1fc551039f1f7f0e4b2e6e4585b60fcfc24370afcf7bd23c6eee\": rpc error: code = NotFound desc = could not find container \"ac73ff1ef41d1fc551039f1f7f0e4b2e6e4585b60fcfc24370afcf7bd23c6eee\": container with ID starting with ac73ff1ef41d1fc551039f1f7f0e4b2e6e4585b60fcfc24370afcf7bd23c6eee not found: ID does not exist" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.092424 4726 scope.go:117] "RemoveContainer" containerID="b5dd4bff78f94c2869b76422340c3a24fae303e59d0a94380cb71ba4c7ea76d1" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.103548 4726 scope.go:117] "RemoveContainer" containerID="6c25c118dc218e081d4ecb0749c39420b1ab543ac3a54a839dc7e28226af14cd" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.116841 4726 scope.go:117] "RemoveContainer" containerID="f8889a56d11176b7ea0555ddcbe3eeafb4287e5c8c3dc141716232345a3a4ec7" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.163315 4726 scope.go:117] "RemoveContainer" containerID="b5dd4bff78f94c2869b76422340c3a24fae303e59d0a94380cb71ba4c7ea76d1" Oct 01 16:04:02 crc kubenswrapper[4726]: E1001 16:04:02.163817 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5dd4bff78f94c2869b76422340c3a24fae303e59d0a94380cb71ba4c7ea76d1\": container with ID starting with b5dd4bff78f94c2869b76422340c3a24fae303e59d0a94380cb71ba4c7ea76d1 not found: ID does not exist" containerID="b5dd4bff78f94c2869b76422340c3a24fae303e59d0a94380cb71ba4c7ea76d1" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.163845 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5dd4bff78f94c2869b76422340c3a24fae303e59d0a94380cb71ba4c7ea76d1"} err="failed to get container status \"b5dd4bff78f94c2869b76422340c3a24fae303e59d0a94380cb71ba4c7ea76d1\": rpc error: code = NotFound desc = could not find container \"b5dd4bff78f94c2869b76422340c3a24fae303e59d0a94380cb71ba4c7ea76d1\": container with ID starting with b5dd4bff78f94c2869b76422340c3a24fae303e59d0a94380cb71ba4c7ea76d1 not found: ID does not exist" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.163868 4726 scope.go:117] "RemoveContainer" containerID="6c25c118dc218e081d4ecb0749c39420b1ab543ac3a54a839dc7e28226af14cd" Oct 01 16:04:02 crc kubenswrapper[4726]: E1001 16:04:02.164375 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c25c118dc218e081d4ecb0749c39420b1ab543ac3a54a839dc7e28226af14cd\": container with ID starting with 6c25c118dc218e081d4ecb0749c39420b1ab543ac3a54a839dc7e28226af14cd not found: ID does not exist" containerID="6c25c118dc218e081d4ecb0749c39420b1ab543ac3a54a839dc7e28226af14cd" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.164411 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c25c118dc218e081d4ecb0749c39420b1ab543ac3a54a839dc7e28226af14cd"} err="failed to get container status \"6c25c118dc218e081d4ecb0749c39420b1ab543ac3a54a839dc7e28226af14cd\": rpc error: code = NotFound desc = could not find container \"6c25c118dc218e081d4ecb0749c39420b1ab543ac3a54a839dc7e28226af14cd\": container with ID starting with 6c25c118dc218e081d4ecb0749c39420b1ab543ac3a54a839dc7e28226af14cd not found: ID does not exist" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.164431 4726 scope.go:117] "RemoveContainer" containerID="f8889a56d11176b7ea0555ddcbe3eeafb4287e5c8c3dc141716232345a3a4ec7" Oct 01 16:04:02 crc kubenswrapper[4726]: E1001 16:04:02.164779 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8889a56d11176b7ea0555ddcbe3eeafb4287e5c8c3dc141716232345a3a4ec7\": container with ID starting with f8889a56d11176b7ea0555ddcbe3eeafb4287e5c8c3dc141716232345a3a4ec7 not found: ID does not exist" containerID="f8889a56d11176b7ea0555ddcbe3eeafb4287e5c8c3dc141716232345a3a4ec7" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.164808 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8889a56d11176b7ea0555ddcbe3eeafb4287e5c8c3dc141716232345a3a4ec7"} err="failed to get container status \"f8889a56d11176b7ea0555ddcbe3eeafb4287e5c8c3dc141716232345a3a4ec7\": rpc error: code = NotFound desc = could not find container \"f8889a56d11176b7ea0555ddcbe3eeafb4287e5c8c3dc141716232345a3a4ec7\": container with ID starting with f8889a56d11176b7ea0555ddcbe3eeafb4287e5c8c3dc141716232345a3a4ec7 not found: ID does not exist" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.164825 4726 scope.go:117] "RemoveContainer" containerID="9eea3a9c7c3f01e9f507b863a0763605eea5993b9a38e261c9b93eda402923b9" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.177912 4726 scope.go:117] "RemoveContainer" containerID="36d40cc9d96a1451565374389d7888507abb608901063d38197751ff3b8dc079" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.195714 4726 scope.go:117] "RemoveContainer" containerID="ae0edf0ab9f075f4124b0fa4a9107804571ba9f43860034ad1ee2e9283f275ee" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.211407 4726 scope.go:117] "RemoveContainer" containerID="9eea3a9c7c3f01e9f507b863a0763605eea5993b9a38e261c9b93eda402923b9" Oct 01 16:04:02 crc kubenswrapper[4726]: E1001 16:04:02.211716 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9eea3a9c7c3f01e9f507b863a0763605eea5993b9a38e261c9b93eda402923b9\": container with ID starting with 9eea3a9c7c3f01e9f507b863a0763605eea5993b9a38e261c9b93eda402923b9 not found: ID does not exist" containerID="9eea3a9c7c3f01e9f507b863a0763605eea5993b9a38e261c9b93eda402923b9" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.211749 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eea3a9c7c3f01e9f507b863a0763605eea5993b9a38e261c9b93eda402923b9"} err="failed to get container status \"9eea3a9c7c3f01e9f507b863a0763605eea5993b9a38e261c9b93eda402923b9\": rpc error: code = NotFound desc = could not find container \"9eea3a9c7c3f01e9f507b863a0763605eea5993b9a38e261c9b93eda402923b9\": container with ID starting with 9eea3a9c7c3f01e9f507b863a0763605eea5993b9a38e261c9b93eda402923b9 not found: ID does not exist" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.211777 4726 scope.go:117] "RemoveContainer" containerID="36d40cc9d96a1451565374389d7888507abb608901063d38197751ff3b8dc079" Oct 01 16:04:02 crc kubenswrapper[4726]: E1001 16:04:02.212155 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36d40cc9d96a1451565374389d7888507abb608901063d38197751ff3b8dc079\": container with ID starting with 36d40cc9d96a1451565374389d7888507abb608901063d38197751ff3b8dc079 not found: ID does not exist" containerID="36d40cc9d96a1451565374389d7888507abb608901063d38197751ff3b8dc079" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.212182 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36d40cc9d96a1451565374389d7888507abb608901063d38197751ff3b8dc079"} err="failed to get container status \"36d40cc9d96a1451565374389d7888507abb608901063d38197751ff3b8dc079\": rpc error: code = NotFound desc = could not find container \"36d40cc9d96a1451565374389d7888507abb608901063d38197751ff3b8dc079\": container with ID starting with 36d40cc9d96a1451565374389d7888507abb608901063d38197751ff3b8dc079 not found: ID does not exist" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.212200 4726 scope.go:117] "RemoveContainer" containerID="ae0edf0ab9f075f4124b0fa4a9107804571ba9f43860034ad1ee2e9283f275ee" Oct 01 16:04:02 crc kubenswrapper[4726]: E1001 16:04:02.212391 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae0edf0ab9f075f4124b0fa4a9107804571ba9f43860034ad1ee2e9283f275ee\": container with ID starting with ae0edf0ab9f075f4124b0fa4a9107804571ba9f43860034ad1ee2e9283f275ee not found: ID does not exist" containerID="ae0edf0ab9f075f4124b0fa4a9107804571ba9f43860034ad1ee2e9283f275ee" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.212413 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae0edf0ab9f075f4124b0fa4a9107804571ba9f43860034ad1ee2e9283f275ee"} err="failed to get container status \"ae0edf0ab9f075f4124b0fa4a9107804571ba9f43860034ad1ee2e9283f275ee\": rpc error: code = NotFound desc = could not find container \"ae0edf0ab9f075f4124b0fa4a9107804571ba9f43860034ad1ee2e9283f275ee\": container with ID starting with ae0edf0ab9f075f4124b0fa4a9107804571ba9f43860034ad1ee2e9283f275ee not found: ID does not exist" Oct 01 16:04:02 crc kubenswrapper[4726]: I1001 16:04:02.993214 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-g69q5" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.200399 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ms6sz"] Oct 01 16:04:03 crc kubenswrapper[4726]: E1001 16:04:03.200828 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" containerName="extract-content" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.200843 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" containerName="extract-content" Oct 01 16:04:03 crc kubenswrapper[4726]: E1001 16:04:03.200853 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6406156d-358e-4c24-9e26-921d117dc024" containerName="extract-utilities" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.200861 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6406156d-358e-4c24-9e26-921d117dc024" containerName="extract-utilities" Oct 01 16:04:03 crc kubenswrapper[4726]: E1001 16:04:03.200872 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b53fe26-735b-47b1-b772-8de0898560ee" containerName="extract-utilities" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.200879 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b53fe26-735b-47b1-b772-8de0898560ee" containerName="extract-utilities" Oct 01 16:04:03 crc kubenswrapper[4726]: E1001 16:04:03.200887 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a7a60ae-9178-467a-b52e-8e9c37dc64a3" containerName="extract-utilities" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.200894 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a7a60ae-9178-467a-b52e-8e9c37dc64a3" containerName="extract-utilities" Oct 01 16:04:03 crc kubenswrapper[4726]: E1001 16:04:03.200902 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a7a60ae-9178-467a-b52e-8e9c37dc64a3" containerName="extract-content" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.200910 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a7a60ae-9178-467a-b52e-8e9c37dc64a3" containerName="extract-content" Oct 01 16:04:03 crc kubenswrapper[4726]: E1001 16:04:03.200920 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" containerName="registry-server" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.200928 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" containerName="registry-server" Oct 01 16:04:03 crc kubenswrapper[4726]: E1001 16:04:03.200939 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" containerName="extract-utilities" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.200947 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" containerName="extract-utilities" Oct 01 16:04:03 crc kubenswrapper[4726]: E1001 16:04:03.200959 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a7a60ae-9178-467a-b52e-8e9c37dc64a3" containerName="registry-server" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.200969 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a7a60ae-9178-467a-b52e-8e9c37dc64a3" containerName="registry-server" Oct 01 16:04:03 crc kubenswrapper[4726]: E1001 16:04:03.200980 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6406156d-358e-4c24-9e26-921d117dc024" containerName="extract-content" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.200988 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6406156d-358e-4c24-9e26-921d117dc024" containerName="extract-content" Oct 01 16:04:03 crc kubenswrapper[4726]: E1001 16:04:03.200998 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b53fe26-735b-47b1-b772-8de0898560ee" containerName="registry-server" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.201005 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b53fe26-735b-47b1-b772-8de0898560ee" containerName="registry-server" Oct 01 16:04:03 crc kubenswrapper[4726]: E1001 16:04:03.201017 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c442dae3-d66d-47a2-b1dc-30332e947203" containerName="marketplace-operator" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.201025 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c442dae3-d66d-47a2-b1dc-30332e947203" containerName="marketplace-operator" Oct 01 16:04:03 crc kubenswrapper[4726]: E1001 16:04:03.201034 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6406156d-358e-4c24-9e26-921d117dc024" containerName="registry-server" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.201065 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6406156d-358e-4c24-9e26-921d117dc024" containerName="registry-server" Oct 01 16:04:03 crc kubenswrapper[4726]: E1001 16:04:03.201080 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b53fe26-735b-47b1-b772-8de0898560ee" containerName="extract-content" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.201088 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b53fe26-735b-47b1-b772-8de0898560ee" containerName="extract-content" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.201207 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6406156d-358e-4c24-9e26-921d117dc024" containerName="registry-server" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.201223 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" containerName="registry-server" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.201231 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c442dae3-d66d-47a2-b1dc-30332e947203" containerName="marketplace-operator" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.201248 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a7a60ae-9178-467a-b52e-8e9c37dc64a3" containerName="registry-server" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.201257 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b53fe26-735b-47b1-b772-8de0898560ee" containerName="registry-server" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.202413 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.204630 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.214121 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ms6sz"] Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.269651 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlf24\" (UniqueName: \"kubernetes.io/projected/4753c482-c4a6-44e9-ace2-7997ee15d241-kube-api-access-nlf24\") pod \"redhat-marketplace-ms6sz\" (UID: \"4753c482-c4a6-44e9-ace2-7997ee15d241\") " pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.269699 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4753c482-c4a6-44e9-ace2-7997ee15d241-utilities\") pod \"redhat-marketplace-ms6sz\" (UID: \"4753c482-c4a6-44e9-ace2-7997ee15d241\") " pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.269830 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4753c482-c4a6-44e9-ace2-7997ee15d241-catalog-content\") pod \"redhat-marketplace-ms6sz\" (UID: \"4753c482-c4a6-44e9-ace2-7997ee15d241\") " pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.370705 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlf24\" (UniqueName: \"kubernetes.io/projected/4753c482-c4a6-44e9-ace2-7997ee15d241-kube-api-access-nlf24\") pod \"redhat-marketplace-ms6sz\" (UID: \"4753c482-c4a6-44e9-ace2-7997ee15d241\") " pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.370739 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4753c482-c4a6-44e9-ace2-7997ee15d241-utilities\") pod \"redhat-marketplace-ms6sz\" (UID: \"4753c482-c4a6-44e9-ace2-7997ee15d241\") " pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.370776 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4753c482-c4a6-44e9-ace2-7997ee15d241-catalog-content\") pod \"redhat-marketplace-ms6sz\" (UID: \"4753c482-c4a6-44e9-ace2-7997ee15d241\") " pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.371220 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4753c482-c4a6-44e9-ace2-7997ee15d241-catalog-content\") pod \"redhat-marketplace-ms6sz\" (UID: \"4753c482-c4a6-44e9-ace2-7997ee15d241\") " pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.371239 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4753c482-c4a6-44e9-ace2-7997ee15d241-utilities\") pod \"redhat-marketplace-ms6sz\" (UID: \"4753c482-c4a6-44e9-ace2-7997ee15d241\") " pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.406205 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ch575"] Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.414096 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.416407 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.417786 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlf24\" (UniqueName: \"kubernetes.io/projected/4753c482-c4a6-44e9-ace2-7997ee15d241-kube-api-access-nlf24\") pod \"redhat-marketplace-ms6sz\" (UID: \"4753c482-c4a6-44e9-ace2-7997ee15d241\") " pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.430096 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ch575"] Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.530914 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.573497 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/039da856-2b47-4aff-b18b-9e9cff090200-catalog-content\") pod \"community-operators-ch575\" (UID: \"039da856-2b47-4aff-b18b-9e9cff090200\") " pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.573852 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdlhx\" (UniqueName: \"kubernetes.io/projected/039da856-2b47-4aff-b18b-9e9cff090200-kube-api-access-wdlhx\") pod \"community-operators-ch575\" (UID: \"039da856-2b47-4aff-b18b-9e9cff090200\") " pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.573882 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/039da856-2b47-4aff-b18b-9e9cff090200-utilities\") pod \"community-operators-ch575\" (UID: \"039da856-2b47-4aff-b18b-9e9cff090200\") " pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.674938 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdlhx\" (UniqueName: \"kubernetes.io/projected/039da856-2b47-4aff-b18b-9e9cff090200-kube-api-access-wdlhx\") pod \"community-operators-ch575\" (UID: \"039da856-2b47-4aff-b18b-9e9cff090200\") " pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.674978 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/039da856-2b47-4aff-b18b-9e9cff090200-utilities\") pod \"community-operators-ch575\" (UID: \"039da856-2b47-4aff-b18b-9e9cff090200\") " pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.675023 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/039da856-2b47-4aff-b18b-9e9cff090200-catalog-content\") pod \"community-operators-ch575\" (UID: \"039da856-2b47-4aff-b18b-9e9cff090200\") " pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.675562 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/039da856-2b47-4aff-b18b-9e9cff090200-catalog-content\") pod \"community-operators-ch575\" (UID: \"039da856-2b47-4aff-b18b-9e9cff090200\") " pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.676249 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/039da856-2b47-4aff-b18b-9e9cff090200-utilities\") pod \"community-operators-ch575\" (UID: \"039da856-2b47-4aff-b18b-9e9cff090200\") " pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.696791 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdlhx\" (UniqueName: \"kubernetes.io/projected/039da856-2b47-4aff-b18b-9e9cff090200-kube-api-access-wdlhx\") pod \"community-operators-ch575\" (UID: \"039da856-2b47-4aff-b18b-9e9cff090200\") " pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.730571 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ms6sz"] Oct 01 16:04:03 crc kubenswrapper[4726]: W1001 16:04:03.738155 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4753c482_c4a6_44e9_ace2_7997ee15d241.slice/crio-4ae919411587640a41ddc3b82cae22c9ae6afe42ade958047ff5ec1c3abdba55 WatchSource:0}: Error finding container 4ae919411587640a41ddc3b82cae22c9ae6afe42ade958047ff5ec1c3abdba55: Status 404 returned error can't find the container with id 4ae919411587640a41ddc3b82cae22c9ae6afe42ade958047ff5ec1c3abdba55 Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.749671 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.830270 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6406156d-358e-4c24-9e26-921d117dc024" path="/var/lib/kubelet/pods/6406156d-358e-4c24-9e26-921d117dc024/volumes" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.832554 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dae1d83-2eb4-4f56-b14d-1d46fd6908eb" path="/var/lib/kubelet/pods/6dae1d83-2eb4-4f56-b14d-1d46fd6908eb/volumes" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.834641 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a7a60ae-9178-467a-b52e-8e9c37dc64a3" path="/var/lib/kubelet/pods/9a7a60ae-9178-467a-b52e-8e9c37dc64a3/volumes" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.836403 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b53fe26-735b-47b1-b772-8de0898560ee" path="/var/lib/kubelet/pods/9b53fe26-735b-47b1-b772-8de0898560ee/volumes" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.837272 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c442dae3-d66d-47a2-b1dc-30332e947203" path="/var/lib/kubelet/pods/c442dae3-d66d-47a2-b1dc-30332e947203/volumes" Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.920940 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ch575"] Oct 01 16:04:03 crc kubenswrapper[4726]: W1001 16:04:03.947594 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod039da856_2b47_4aff_b18b_9e9cff090200.slice/crio-7224e3d0cb74894a8075d5dc6a02df5bddf74d748bbd58e4a176ff634609d570 WatchSource:0}: Error finding container 7224e3d0cb74894a8075d5dc6a02df5bddf74d748bbd58e4a176ff634609d570: Status 404 returned error can't find the container with id 7224e3d0cb74894a8075d5dc6a02df5bddf74d748bbd58e4a176ff634609d570 Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.997385 4726 generic.go:334] "Generic (PLEG): container finished" podID="4753c482-c4a6-44e9-ace2-7997ee15d241" containerID="7c16945040a335caf2ff70c5f058e2dfcca001eb6a92cefec7eadb33609e1458" exitCode=0 Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.997464 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ms6sz" event={"ID":"4753c482-c4a6-44e9-ace2-7997ee15d241","Type":"ContainerDied","Data":"7c16945040a335caf2ff70c5f058e2dfcca001eb6a92cefec7eadb33609e1458"} Oct 01 16:04:03 crc kubenswrapper[4726]: I1001 16:04:03.997495 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ms6sz" event={"ID":"4753c482-c4a6-44e9-ace2-7997ee15d241","Type":"ContainerStarted","Data":"4ae919411587640a41ddc3b82cae22c9ae6afe42ade958047ff5ec1c3abdba55"} Oct 01 16:04:04 crc kubenswrapper[4726]: I1001 16:04:04.000392 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ch575" event={"ID":"039da856-2b47-4aff-b18b-9e9cff090200","Type":"ContainerStarted","Data":"7224e3d0cb74894a8075d5dc6a02df5bddf74d748bbd58e4a176ff634609d570"} Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.005927 4726 generic.go:334] "Generic (PLEG): container finished" podID="039da856-2b47-4aff-b18b-9e9cff090200" containerID="3ad839ddf849aef0de6c2d68e3ebc1a2b09a8953f3c79ceea510b824256a36aa" exitCode=0 Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.006023 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ch575" event={"ID":"039da856-2b47-4aff-b18b-9e9cff090200","Type":"ContainerDied","Data":"3ad839ddf849aef0de6c2d68e3ebc1a2b09a8953f3c79ceea510b824256a36aa"} Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.009119 4726 generic.go:334] "Generic (PLEG): container finished" podID="4753c482-c4a6-44e9-ace2-7997ee15d241" containerID="6166ef2e6475575f866aa3d42650c5f51040db77c6f77966612f353069458b86" exitCode=0 Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.009168 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ms6sz" event={"ID":"4753c482-c4a6-44e9-ace2-7997ee15d241","Type":"ContainerDied","Data":"6166ef2e6475575f866aa3d42650c5f51040db77c6f77966612f353069458b86"} Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.610982 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r5rkm"] Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.612686 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.616898 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.630883 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r5rkm"] Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.797671 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2ql9b"] Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.807169 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dbcbacf-6b7a-4747-a278-021abf5fd244-utilities\") pod \"certified-operators-r5rkm\" (UID: \"6dbcbacf-6b7a-4747-a278-021abf5fd244\") " pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.807217 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dbcbacf-6b7a-4747-a278-021abf5fd244-catalog-content\") pod \"certified-operators-r5rkm\" (UID: \"6dbcbacf-6b7a-4747-a278-021abf5fd244\") " pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.807258 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26zgg\" (UniqueName: \"kubernetes.io/projected/6dbcbacf-6b7a-4747-a278-021abf5fd244-kube-api-access-26zgg\") pod \"certified-operators-r5rkm\" (UID: \"6dbcbacf-6b7a-4747-a278-021abf5fd244\") " pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.807333 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.809114 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.817179 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2ql9b"] Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.908372 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26zgg\" (UniqueName: \"kubernetes.io/projected/6dbcbacf-6b7a-4747-a278-021abf5fd244-kube-api-access-26zgg\") pod \"certified-operators-r5rkm\" (UID: \"6dbcbacf-6b7a-4747-a278-021abf5fd244\") " pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.908432 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-catalog-content\") pod \"redhat-operators-2ql9b\" (UID: \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\") " pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.908459 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc5lz\" (UniqueName: \"kubernetes.io/projected/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-kube-api-access-nc5lz\") pod \"redhat-operators-2ql9b\" (UID: \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\") " pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.908484 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-utilities\") pod \"redhat-operators-2ql9b\" (UID: \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\") " pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.908526 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dbcbacf-6b7a-4747-a278-021abf5fd244-utilities\") pod \"certified-operators-r5rkm\" (UID: \"6dbcbacf-6b7a-4747-a278-021abf5fd244\") " pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.908544 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dbcbacf-6b7a-4747-a278-021abf5fd244-catalog-content\") pod \"certified-operators-r5rkm\" (UID: \"6dbcbacf-6b7a-4747-a278-021abf5fd244\") " pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.908916 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dbcbacf-6b7a-4747-a278-021abf5fd244-catalog-content\") pod \"certified-operators-r5rkm\" (UID: \"6dbcbacf-6b7a-4747-a278-021abf5fd244\") " pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.909443 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dbcbacf-6b7a-4747-a278-021abf5fd244-utilities\") pod \"certified-operators-r5rkm\" (UID: \"6dbcbacf-6b7a-4747-a278-021abf5fd244\") " pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.930216 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26zgg\" (UniqueName: \"kubernetes.io/projected/6dbcbacf-6b7a-4747-a278-021abf5fd244-kube-api-access-26zgg\") pod \"certified-operators-r5rkm\" (UID: \"6dbcbacf-6b7a-4747-a278-021abf5fd244\") " pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:05 crc kubenswrapper[4726]: I1001 16:04:05.945583 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:06 crc kubenswrapper[4726]: I1001 16:04:06.010205 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-catalog-content\") pod \"redhat-operators-2ql9b\" (UID: \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\") " pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:04:06 crc kubenswrapper[4726]: I1001 16:04:06.010262 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc5lz\" (UniqueName: \"kubernetes.io/projected/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-kube-api-access-nc5lz\") pod \"redhat-operators-2ql9b\" (UID: \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\") " pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:04:06 crc kubenswrapper[4726]: I1001 16:04:06.010308 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-utilities\") pod \"redhat-operators-2ql9b\" (UID: \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\") " pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:04:06 crc kubenswrapper[4726]: I1001 16:04:06.010706 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-catalog-content\") pod \"redhat-operators-2ql9b\" (UID: \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\") " pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:04:06 crc kubenswrapper[4726]: I1001 16:04:06.010768 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-utilities\") pod \"redhat-operators-2ql9b\" (UID: \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\") " pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:04:06 crc kubenswrapper[4726]: I1001 16:04:06.017457 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ch575" event={"ID":"039da856-2b47-4aff-b18b-9e9cff090200","Type":"ContainerStarted","Data":"534f0cfbb11a2af4819b017d5f8b4b4cf20a5f4c5d2535491542290a1df72d6f"} Oct 01 16:04:06 crc kubenswrapper[4726]: I1001 16:04:06.031596 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ms6sz" event={"ID":"4753c482-c4a6-44e9-ace2-7997ee15d241","Type":"ContainerStarted","Data":"8611784edd4fb1222d077b56e2111f9acf5fcaeff0dccc23a5d9b47856692621"} Oct 01 16:04:06 crc kubenswrapper[4726]: I1001 16:04:06.037855 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc5lz\" (UniqueName: \"kubernetes.io/projected/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-kube-api-access-nc5lz\") pod \"redhat-operators-2ql9b\" (UID: \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\") " pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:04:06 crc kubenswrapper[4726]: I1001 16:04:06.132307 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ms6sz" podStartSLOduration=1.549389185 podStartE2EDuration="3.132283912s" podCreationTimestamp="2025-10-01 16:04:03 +0000 UTC" firstStartedPulling="2025-10-01 16:04:03.999870144 +0000 UTC m=+296.901422731" lastFinishedPulling="2025-10-01 16:04:05.582764871 +0000 UTC m=+298.484317458" observedRunningTime="2025-10-01 16:04:06.06447612 +0000 UTC m=+298.966028697" watchObservedRunningTime="2025-10-01 16:04:06.132283912 +0000 UTC m=+299.033836489" Oct 01 16:04:06 crc kubenswrapper[4726]: I1001 16:04:06.134012 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r5rkm"] Oct 01 16:04:06 crc kubenswrapper[4726]: I1001 16:04:06.134323 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:04:06 crc kubenswrapper[4726]: I1001 16:04:06.531145 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2ql9b"] Oct 01 16:04:07 crc kubenswrapper[4726]: I1001 16:04:07.039212 4726 generic.go:334] "Generic (PLEG): container finished" podID="2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" containerID="4f417a88ac4ad829e3698306fea3f478db1b51b95b6d4d23b5c2ea298f092f7b" exitCode=0 Oct 01 16:04:07 crc kubenswrapper[4726]: I1001 16:04:07.039345 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2ql9b" event={"ID":"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a","Type":"ContainerDied","Data":"4f417a88ac4ad829e3698306fea3f478db1b51b95b6d4d23b5c2ea298f092f7b"} Oct 01 16:04:07 crc kubenswrapper[4726]: I1001 16:04:07.039872 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2ql9b" event={"ID":"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a","Type":"ContainerStarted","Data":"3d6a038380a7e43a6b91a88d289c58817ffbca92845046f646b9578ab1be4fd9"} Oct 01 16:04:07 crc kubenswrapper[4726]: I1001 16:04:07.043713 4726 generic.go:334] "Generic (PLEG): container finished" podID="039da856-2b47-4aff-b18b-9e9cff090200" containerID="534f0cfbb11a2af4819b017d5f8b4b4cf20a5f4c5d2535491542290a1df72d6f" exitCode=0 Oct 01 16:04:07 crc kubenswrapper[4726]: I1001 16:04:07.043885 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ch575" event={"ID":"039da856-2b47-4aff-b18b-9e9cff090200","Type":"ContainerDied","Data":"534f0cfbb11a2af4819b017d5f8b4b4cf20a5f4c5d2535491542290a1df72d6f"} Oct 01 16:04:07 crc kubenswrapper[4726]: I1001 16:04:07.049849 4726 generic.go:334] "Generic (PLEG): container finished" podID="6dbcbacf-6b7a-4747-a278-021abf5fd244" containerID="3c9278e0b3673e86645af6f1c2cf8b4a60871ce7282811caaff41ad16f807cf9" exitCode=0 Oct 01 16:04:07 crc kubenswrapper[4726]: I1001 16:04:07.049908 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5rkm" event={"ID":"6dbcbacf-6b7a-4747-a278-021abf5fd244","Type":"ContainerDied","Data":"3c9278e0b3673e86645af6f1c2cf8b4a60871ce7282811caaff41ad16f807cf9"} Oct 01 16:04:07 crc kubenswrapper[4726]: I1001 16:04:07.049945 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5rkm" event={"ID":"6dbcbacf-6b7a-4747-a278-021abf5fd244","Type":"ContainerStarted","Data":"9242e4e8aa443e780eaa9e6dbb93809659849c7dbfd397f7fac9654082a6102c"} Oct 01 16:04:08 crc kubenswrapper[4726]: I1001 16:04:08.057692 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5rkm" event={"ID":"6dbcbacf-6b7a-4747-a278-021abf5fd244","Type":"ContainerStarted","Data":"f542bbbfa187bb76d9aed6b04285cb8b93acf8bcb87f7a945c41e00927ccd02d"} Oct 01 16:04:08 crc kubenswrapper[4726]: I1001 16:04:08.060553 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ch575" event={"ID":"039da856-2b47-4aff-b18b-9e9cff090200","Type":"ContainerStarted","Data":"58d532d65f1ad28460ea5b227763c39e6b1227b1fe7cdc025de99d3d096324f1"} Oct 01 16:04:08 crc kubenswrapper[4726]: I1001 16:04:08.113000 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ch575" podStartSLOduration=2.600728362 podStartE2EDuration="5.11298162s" podCreationTimestamp="2025-10-01 16:04:03 +0000 UTC" firstStartedPulling="2025-10-01 16:04:05.007062261 +0000 UTC m=+297.908614838" lastFinishedPulling="2025-10-01 16:04:07.519315479 +0000 UTC m=+300.420868096" observedRunningTime="2025-10-01 16:04:08.110447553 +0000 UTC m=+301.012000150" watchObservedRunningTime="2025-10-01 16:04:08.11298162 +0000 UTC m=+301.014534187" Oct 01 16:04:09 crc kubenswrapper[4726]: I1001 16:04:09.066511 4726 generic.go:334] "Generic (PLEG): container finished" podID="2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" containerID="fa1c4c8f3a84a3d7f2be578564919beda38dd35557284336a8c2d505a8cdc82b" exitCode=0 Oct 01 16:04:09 crc kubenswrapper[4726]: I1001 16:04:09.066566 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2ql9b" event={"ID":"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a","Type":"ContainerDied","Data":"fa1c4c8f3a84a3d7f2be578564919beda38dd35557284336a8c2d505a8cdc82b"} Oct 01 16:04:09 crc kubenswrapper[4726]: I1001 16:04:09.072433 4726 generic.go:334] "Generic (PLEG): container finished" podID="6dbcbacf-6b7a-4747-a278-021abf5fd244" containerID="f542bbbfa187bb76d9aed6b04285cb8b93acf8bcb87f7a945c41e00927ccd02d" exitCode=0 Oct 01 16:04:09 crc kubenswrapper[4726]: I1001 16:04:09.073383 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5rkm" event={"ID":"6dbcbacf-6b7a-4747-a278-021abf5fd244","Type":"ContainerDied","Data":"f542bbbfa187bb76d9aed6b04285cb8b93acf8bcb87f7a945c41e00927ccd02d"} Oct 01 16:04:10 crc kubenswrapper[4726]: I1001 16:04:10.081778 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2ql9b" event={"ID":"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a","Type":"ContainerStarted","Data":"a9c9ec66f518e18cec5e6c038d9b0a6ab8e4f31e36c0ad2368da7ce0f6f975a9"} Oct 01 16:04:10 crc kubenswrapper[4726]: I1001 16:04:10.084262 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5rkm" event={"ID":"6dbcbacf-6b7a-4747-a278-021abf5fd244","Type":"ContainerStarted","Data":"f6a73c09add3ddf20d9f298825cf00639945956c574e2f8e8438416311e904f2"} Oct 01 16:04:10 crc kubenswrapper[4726]: I1001 16:04:10.099632 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2ql9b" podStartSLOduration=2.608699371 podStartE2EDuration="5.099602007s" podCreationTimestamp="2025-10-01 16:04:05 +0000 UTC" firstStartedPulling="2025-10-01 16:04:07.041121975 +0000 UTC m=+299.942674592" lastFinishedPulling="2025-10-01 16:04:09.532024651 +0000 UTC m=+302.433577228" observedRunningTime="2025-10-01 16:04:10.098458992 +0000 UTC m=+303.000011569" watchObservedRunningTime="2025-10-01 16:04:10.099602007 +0000 UTC m=+303.001154584" Oct 01 16:04:10 crc kubenswrapper[4726]: I1001 16:04:10.122450 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r5rkm" podStartSLOduration=2.5567319360000003 podStartE2EDuration="5.122426574s" podCreationTimestamp="2025-10-01 16:04:05 +0000 UTC" firstStartedPulling="2025-10-01 16:04:07.05155935 +0000 UTC m=+299.953111937" lastFinishedPulling="2025-10-01 16:04:09.617253998 +0000 UTC m=+302.518806575" observedRunningTime="2025-10-01 16:04:10.119689482 +0000 UTC m=+303.021242079" watchObservedRunningTime="2025-10-01 16:04:10.122426574 +0000 UTC m=+303.023979161" Oct 01 16:04:13 crc kubenswrapper[4726]: I1001 16:04:13.531401 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:13 crc kubenswrapper[4726]: I1001 16:04:13.531917 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:13 crc kubenswrapper[4726]: I1001 16:04:13.588603 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:13 crc kubenswrapper[4726]: I1001 16:04:13.750507 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:13 crc kubenswrapper[4726]: I1001 16:04:13.750839 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:13 crc kubenswrapper[4726]: I1001 16:04:13.794427 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:14 crc kubenswrapper[4726]: I1001 16:04:14.150869 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ms6sz" Oct 01 16:04:14 crc kubenswrapper[4726]: I1001 16:04:14.151092 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ch575" Oct 01 16:04:15 crc kubenswrapper[4726]: I1001 16:04:15.946533 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:15 crc kubenswrapper[4726]: I1001 16:04:15.947467 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:15 crc kubenswrapper[4726]: I1001 16:04:15.992323 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:16 crc kubenswrapper[4726]: I1001 16:04:16.135159 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:04:16 crc kubenswrapper[4726]: I1001 16:04:16.135195 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:04:16 crc kubenswrapper[4726]: I1001 16:04:16.161436 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:04:16 crc kubenswrapper[4726]: I1001 16:04:16.175192 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:04:17 crc kubenswrapper[4726]: I1001 16:04:17.171566 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:05:23 crc kubenswrapper[4726]: I1001 16:05:23.414406 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:05:23 crc kubenswrapper[4726]: I1001 16:05:23.415255 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:05:53 crc kubenswrapper[4726]: I1001 16:05:53.414283 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:05:53 crc kubenswrapper[4726]: I1001 16:05:53.414934 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:06:23 crc kubenswrapper[4726]: I1001 16:06:23.413831 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:06:23 crc kubenswrapper[4726]: I1001 16:06:23.414428 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:06:23 crc kubenswrapper[4726]: I1001 16:06:23.414491 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 16:06:23 crc kubenswrapper[4726]: I1001 16:06:23.415473 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"29994399a03c1d7203fa51b4be7f3ba156ce8c41237e36d940e730558f1a7456"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:06:23 crc kubenswrapper[4726]: I1001 16:06:23.415602 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://29994399a03c1d7203fa51b4be7f3ba156ce8c41237e36d940e730558f1a7456" gracePeriod=600 Oct 01 16:06:23 crc kubenswrapper[4726]: I1001 16:06:23.910752 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="29994399a03c1d7203fa51b4be7f3ba156ce8c41237e36d940e730558f1a7456" exitCode=0 Oct 01 16:06:23 crc kubenswrapper[4726]: I1001 16:06:23.910788 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"29994399a03c1d7203fa51b4be7f3ba156ce8c41237e36d940e730558f1a7456"} Oct 01 16:06:23 crc kubenswrapper[4726]: I1001 16:06:23.911340 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"d8eaf0f51eb5e70b61a0a405367175f41860b9a554b95b023ac96b6912d9a4f7"} Oct 01 16:06:23 crc kubenswrapper[4726]: I1001 16:06:23.911410 4726 scope.go:117] "RemoveContainer" containerID="5c3c134746d4103f56296e51caa9247f7fbb74a7b43b48d49676173b370d77b6" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.218930 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8jwx8"] Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.220102 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.238677 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8jwx8"] Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.354143 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b519b97e-33e0-42b7-913a-7e95abda2318-registry-tls\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.354200 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l9nj\" (UniqueName: \"kubernetes.io/projected/b519b97e-33e0-42b7-913a-7e95abda2318-kube-api-access-5l9nj\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.354223 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b519b97e-33e0-42b7-913a-7e95abda2318-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.354282 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b519b97e-33e0-42b7-913a-7e95abda2318-bound-sa-token\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.354300 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b519b97e-33e0-42b7-913a-7e95abda2318-registry-certificates\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.354366 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b519b97e-33e0-42b7-913a-7e95abda2318-trusted-ca\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.354389 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b519b97e-33e0-42b7-913a-7e95abda2318-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.354419 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.376268 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.455215 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b519b97e-33e0-42b7-913a-7e95abda2318-registry-tls\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.455441 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l9nj\" (UniqueName: \"kubernetes.io/projected/b519b97e-33e0-42b7-913a-7e95abda2318-kube-api-access-5l9nj\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.455626 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b519b97e-33e0-42b7-913a-7e95abda2318-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.455738 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b519b97e-33e0-42b7-913a-7e95abda2318-bound-sa-token\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.455773 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b519b97e-33e0-42b7-913a-7e95abda2318-registry-certificates\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.455851 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b519b97e-33e0-42b7-913a-7e95abda2318-trusted-ca\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.455909 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b519b97e-33e0-42b7-913a-7e95abda2318-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.456459 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b519b97e-33e0-42b7-913a-7e95abda2318-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.456954 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b519b97e-33e0-42b7-913a-7e95abda2318-trusted-ca\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.457132 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b519b97e-33e0-42b7-913a-7e95abda2318-registry-certificates\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.461718 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b519b97e-33e0-42b7-913a-7e95abda2318-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.461949 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b519b97e-33e0-42b7-913a-7e95abda2318-registry-tls\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.471230 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b519b97e-33e0-42b7-913a-7e95abda2318-bound-sa-token\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.474422 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l9nj\" (UniqueName: \"kubernetes.io/projected/b519b97e-33e0-42b7-913a-7e95abda2318-kube-api-access-5l9nj\") pod \"image-registry-66df7c8f76-8jwx8\" (UID: \"b519b97e-33e0-42b7-913a-7e95abda2318\") " pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.537481 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:41 crc kubenswrapper[4726]: I1001 16:06:41.769732 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8jwx8"] Oct 01 16:06:42 crc kubenswrapper[4726]: I1001 16:06:42.042427 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" event={"ID":"b519b97e-33e0-42b7-913a-7e95abda2318","Type":"ContainerStarted","Data":"a36d9bea5891c3415fb70a923497d96719c197898890d4b12fd8a2e5a9c4f5c6"} Oct 01 16:06:42 crc kubenswrapper[4726]: I1001 16:06:42.042570 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" event={"ID":"b519b97e-33e0-42b7-913a-7e95abda2318","Type":"ContainerStarted","Data":"cb46e87ab64ddbecfd44bb50b31db1087943aff9c163d899a1ee1500ceec0a78"} Oct 01 16:06:42 crc kubenswrapper[4726]: I1001 16:06:42.043027 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:06:42 crc kubenswrapper[4726]: I1001 16:06:42.074755 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" podStartSLOduration=1.074724112 podStartE2EDuration="1.074724112s" podCreationTimestamp="2025-10-01 16:06:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:06:42.068029035 +0000 UTC m=+454.969581642" watchObservedRunningTime="2025-10-01 16:06:42.074724112 +0000 UTC m=+454.976276739" Oct 01 16:07:01 crc kubenswrapper[4726]: I1001 16:07:01.548509 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-8jwx8" Oct 01 16:07:01 crc kubenswrapper[4726]: I1001 16:07:01.613019 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j9fr5"] Oct 01 16:07:26 crc kubenswrapper[4726]: I1001 16:07:26.670550 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" podUID="496e7e53-b6f6-48b1-9657-a091d21c301f" containerName="registry" containerID="cri-o://944ceea21ccc4b37f40718a1146700fe4fb37c6baf6ff823552f3b984b295869" gracePeriod=30 Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.010537 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.172585 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/496e7e53-b6f6-48b1-9657-a091d21c301f-trusted-ca\") pod \"496e7e53-b6f6-48b1-9657-a091d21c301f\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.172697 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/496e7e53-b6f6-48b1-9657-a091d21c301f-registry-certificates\") pod \"496e7e53-b6f6-48b1-9657-a091d21c301f\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.172794 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/496e7e53-b6f6-48b1-9657-a091d21c301f-installation-pull-secrets\") pod \"496e7e53-b6f6-48b1-9657-a091d21c301f\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.172869 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-registry-tls\") pod \"496e7e53-b6f6-48b1-9657-a091d21c301f\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.172934 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwzxw\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-kube-api-access-pwzxw\") pod \"496e7e53-b6f6-48b1-9657-a091d21c301f\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.172983 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/496e7e53-b6f6-48b1-9657-a091d21c301f-ca-trust-extracted\") pod \"496e7e53-b6f6-48b1-9657-a091d21c301f\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.173123 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-bound-sa-token\") pod \"496e7e53-b6f6-48b1-9657-a091d21c301f\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.173444 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"496e7e53-b6f6-48b1-9657-a091d21c301f\" (UID: \"496e7e53-b6f6-48b1-9657-a091d21c301f\") " Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.173643 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e7e53-b6f6-48b1-9657-a091d21c301f-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "496e7e53-b6f6-48b1-9657-a091d21c301f" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.173915 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/496e7e53-b6f6-48b1-9657-a091d21c301f-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.174219 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e7e53-b6f6-48b1-9657-a091d21c301f-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "496e7e53-b6f6-48b1-9657-a091d21c301f" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.179467 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e7e53-b6f6-48b1-9657-a091d21c301f-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "496e7e53-b6f6-48b1-9657-a091d21c301f" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.179568 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "496e7e53-b6f6-48b1-9657-a091d21c301f" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.180287 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-kube-api-access-pwzxw" (OuterVolumeSpecName: "kube-api-access-pwzxw") pod "496e7e53-b6f6-48b1-9657-a091d21c301f" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f"). InnerVolumeSpecName "kube-api-access-pwzxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.184741 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "496e7e53-b6f6-48b1-9657-a091d21c301f" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.191642 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "496e7e53-b6f6-48b1-9657-a091d21c301f" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.198702 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/496e7e53-b6f6-48b1-9657-a091d21c301f-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "496e7e53-b6f6-48b1-9657-a091d21c301f" (UID: "496e7e53-b6f6-48b1-9657-a091d21c301f"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.274707 4726 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.274754 4726 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/496e7e53-b6f6-48b1-9657-a091d21c301f-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.274767 4726 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/496e7e53-b6f6-48b1-9657-a091d21c301f-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.274780 4726 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.274788 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwzxw\" (UniqueName: \"kubernetes.io/projected/496e7e53-b6f6-48b1-9657-a091d21c301f-kube-api-access-pwzxw\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.274797 4726 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/496e7e53-b6f6-48b1-9657-a091d21c301f-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.334677 4726 generic.go:334] "Generic (PLEG): container finished" podID="496e7e53-b6f6-48b1-9657-a091d21c301f" containerID="944ceea21ccc4b37f40718a1146700fe4fb37c6baf6ff823552f3b984b295869" exitCode=0 Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.334728 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" event={"ID":"496e7e53-b6f6-48b1-9657-a091d21c301f","Type":"ContainerDied","Data":"944ceea21ccc4b37f40718a1146700fe4fb37c6baf6ff823552f3b984b295869"} Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.334764 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" event={"ID":"496e7e53-b6f6-48b1-9657-a091d21c301f","Type":"ContainerDied","Data":"19f64afe7167ee0df00b74e33c69ddc2b79077ede2f6bf7de0cd929487b9d597"} Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.334783 4726 scope.go:117] "RemoveContainer" containerID="944ceea21ccc4b37f40718a1146700fe4fb37c6baf6ff823552f3b984b295869" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.334710 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-j9fr5" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.366235 4726 scope.go:117] "RemoveContainer" containerID="944ceea21ccc4b37f40718a1146700fe4fb37c6baf6ff823552f3b984b295869" Oct 01 16:07:27 crc kubenswrapper[4726]: E1001 16:07:27.366833 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"944ceea21ccc4b37f40718a1146700fe4fb37c6baf6ff823552f3b984b295869\": container with ID starting with 944ceea21ccc4b37f40718a1146700fe4fb37c6baf6ff823552f3b984b295869 not found: ID does not exist" containerID="944ceea21ccc4b37f40718a1146700fe4fb37c6baf6ff823552f3b984b295869" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.366969 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"944ceea21ccc4b37f40718a1146700fe4fb37c6baf6ff823552f3b984b295869"} err="failed to get container status \"944ceea21ccc4b37f40718a1146700fe4fb37c6baf6ff823552f3b984b295869\": rpc error: code = NotFound desc = could not find container \"944ceea21ccc4b37f40718a1146700fe4fb37c6baf6ff823552f3b984b295869\": container with ID starting with 944ceea21ccc4b37f40718a1146700fe4fb37c6baf6ff823552f3b984b295869 not found: ID does not exist" Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.369233 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j9fr5"] Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.372821 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j9fr5"] Oct 01 16:07:27 crc kubenswrapper[4726]: I1001 16:07:27.815439 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e7e53-b6f6-48b1-9657-a091d21c301f" path="/var/lib/kubelet/pods/496e7e53-b6f6-48b1-9657-a091d21c301f/volumes" Oct 01 16:08:23 crc kubenswrapper[4726]: I1001 16:08:23.414099 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:08:23 crc kubenswrapper[4726]: I1001 16:08:23.414714 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:08:53 crc kubenswrapper[4726]: I1001 16:08:53.414139 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:08:53 crc kubenswrapper[4726]: I1001 16:08:53.414967 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:09:23 crc kubenswrapper[4726]: I1001 16:09:23.413703 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:09:23 crc kubenswrapper[4726]: I1001 16:09:23.414305 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:09:23 crc kubenswrapper[4726]: I1001 16:09:23.414352 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 16:09:23 crc kubenswrapper[4726]: I1001 16:09:23.415183 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d8eaf0f51eb5e70b61a0a405367175f41860b9a554b95b023ac96b6912d9a4f7"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:09:23 crc kubenswrapper[4726]: I1001 16:09:23.415242 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://d8eaf0f51eb5e70b61a0a405367175f41860b9a554b95b023ac96b6912d9a4f7" gracePeriod=600 Oct 01 16:09:24 crc kubenswrapper[4726]: I1001 16:09:24.106900 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="d8eaf0f51eb5e70b61a0a405367175f41860b9a554b95b023ac96b6912d9a4f7" exitCode=0 Oct 01 16:09:24 crc kubenswrapper[4726]: I1001 16:09:24.107378 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"d8eaf0f51eb5e70b61a0a405367175f41860b9a554b95b023ac96b6912d9a4f7"} Oct 01 16:09:24 crc kubenswrapper[4726]: I1001 16:09:24.107583 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"843d4fde9abf12c4ceb01029d43ed7c5445854afa4d6b76b3edc8ef114cd52e4"} Oct 01 16:09:24 crc kubenswrapper[4726]: I1001 16:09:24.107621 4726 scope.go:117] "RemoveContainer" containerID="29994399a03c1d7203fa51b4be7f3ba156ce8c41237e36d940e730558f1a7456" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.062657 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pz8pb"] Oct 01 16:09:28 crc kubenswrapper[4726]: E1001 16:09:28.063409 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="496e7e53-b6f6-48b1-9657-a091d21c301f" containerName="registry" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.063421 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="496e7e53-b6f6-48b1-9657-a091d21c301f" containerName="registry" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.063515 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="496e7e53-b6f6-48b1-9657-a091d21c301f" containerName="registry" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.063843 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-pz8pb" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.066970 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.067588 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.068343 4726 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-5nkq8" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.076409 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-jkfpv"] Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.077674 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-jkfpv" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.081590 4726 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-xr2km" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.091284 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-jxd7m"] Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.091984 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-jxd7m" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.107591 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-jkfpv"] Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.108457 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzbcl\" (UniqueName: \"kubernetes.io/projected/55df3602-b3df-45f4-90b8-eddb1903bf32-kube-api-access-rzbcl\") pod \"cert-manager-cainjector-7f985d654d-pz8pb\" (UID: \"55df3602-b3df-45f4-90b8-eddb1903bf32\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pz8pb" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.111687 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj5qt\" (UniqueName: \"kubernetes.io/projected/cfc89425-6bca-4267-9a51-50e0d8325eaa-kube-api-access-xj5qt\") pod \"cert-manager-5b446d88c5-jkfpv\" (UID: \"cfc89425-6bca-4267-9a51-50e0d8325eaa\") " pod="cert-manager/cert-manager-5b446d88c5-jkfpv" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.113224 4726 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-n4ssz" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.144755 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-jxd7m"] Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.147498 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pz8pb"] Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.214208 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqzq4\" (UniqueName: \"kubernetes.io/projected/51706d0c-886c-4865-ba8b-ade37f39b84e-kube-api-access-bqzq4\") pod \"cert-manager-webhook-5655c58dd6-jxd7m\" (UID: \"51706d0c-886c-4865-ba8b-ade37f39b84e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-jxd7m" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.214280 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj5qt\" (UniqueName: \"kubernetes.io/projected/cfc89425-6bca-4267-9a51-50e0d8325eaa-kube-api-access-xj5qt\") pod \"cert-manager-5b446d88c5-jkfpv\" (UID: \"cfc89425-6bca-4267-9a51-50e0d8325eaa\") " pod="cert-manager/cert-manager-5b446d88c5-jkfpv" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.214360 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzbcl\" (UniqueName: \"kubernetes.io/projected/55df3602-b3df-45f4-90b8-eddb1903bf32-kube-api-access-rzbcl\") pod \"cert-manager-cainjector-7f985d654d-pz8pb\" (UID: \"55df3602-b3df-45f4-90b8-eddb1903bf32\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pz8pb" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.234986 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj5qt\" (UniqueName: \"kubernetes.io/projected/cfc89425-6bca-4267-9a51-50e0d8325eaa-kube-api-access-xj5qt\") pod \"cert-manager-5b446d88c5-jkfpv\" (UID: \"cfc89425-6bca-4267-9a51-50e0d8325eaa\") " pod="cert-manager/cert-manager-5b446d88c5-jkfpv" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.238345 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzbcl\" (UniqueName: \"kubernetes.io/projected/55df3602-b3df-45f4-90b8-eddb1903bf32-kube-api-access-rzbcl\") pod \"cert-manager-cainjector-7f985d654d-pz8pb\" (UID: \"55df3602-b3df-45f4-90b8-eddb1903bf32\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pz8pb" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.315623 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqzq4\" (UniqueName: \"kubernetes.io/projected/51706d0c-886c-4865-ba8b-ade37f39b84e-kube-api-access-bqzq4\") pod \"cert-manager-webhook-5655c58dd6-jxd7m\" (UID: \"51706d0c-886c-4865-ba8b-ade37f39b84e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-jxd7m" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.338128 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqzq4\" (UniqueName: \"kubernetes.io/projected/51706d0c-886c-4865-ba8b-ade37f39b84e-kube-api-access-bqzq4\") pod \"cert-manager-webhook-5655c58dd6-jxd7m\" (UID: \"51706d0c-886c-4865-ba8b-ade37f39b84e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-jxd7m" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.387150 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-pz8pb" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.400699 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-jkfpv" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.443727 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-jxd7m" Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.634973 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-jkfpv"] Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.640609 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.682151 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pz8pb"] Oct 01 16:09:28 crc kubenswrapper[4726]: W1001 16:09:28.694010 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55df3602_b3df_45f4_90b8_eddb1903bf32.slice/crio-91419d4767e85840a9ebbd5ad01507072a72fcfd1e3933f88fc6ed99346ca323 WatchSource:0}: Error finding container 91419d4767e85840a9ebbd5ad01507072a72fcfd1e3933f88fc6ed99346ca323: Status 404 returned error can't find the container with id 91419d4767e85840a9ebbd5ad01507072a72fcfd1e3933f88fc6ed99346ca323 Oct 01 16:09:28 crc kubenswrapper[4726]: I1001 16:09:28.710182 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-jxd7m"] Oct 01 16:09:28 crc kubenswrapper[4726]: W1001 16:09:28.715894 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51706d0c_886c_4865_ba8b_ade37f39b84e.slice/crio-ae658b79997214e8e6bef0be5ceeba8b267f6fd150a231705917277932851254 WatchSource:0}: Error finding container ae658b79997214e8e6bef0be5ceeba8b267f6fd150a231705917277932851254: Status 404 returned error can't find the container with id ae658b79997214e8e6bef0be5ceeba8b267f6fd150a231705917277932851254 Oct 01 16:09:29 crc kubenswrapper[4726]: I1001 16:09:29.154977 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-pz8pb" event={"ID":"55df3602-b3df-45f4-90b8-eddb1903bf32","Type":"ContainerStarted","Data":"91419d4767e85840a9ebbd5ad01507072a72fcfd1e3933f88fc6ed99346ca323"} Oct 01 16:09:29 crc kubenswrapper[4726]: I1001 16:09:29.156518 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-jxd7m" event={"ID":"51706d0c-886c-4865-ba8b-ade37f39b84e","Type":"ContainerStarted","Data":"ae658b79997214e8e6bef0be5ceeba8b267f6fd150a231705917277932851254"} Oct 01 16:09:29 crc kubenswrapper[4726]: I1001 16:09:29.157767 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-jkfpv" event={"ID":"cfc89425-6bca-4267-9a51-50e0d8325eaa","Type":"ContainerStarted","Data":"610663ff8147c9c76d986af151d951a5e36657ecf3b57be735289f16f5e146a2"} Oct 01 16:09:33 crc kubenswrapper[4726]: I1001 16:09:33.177341 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-jkfpv" event={"ID":"cfc89425-6bca-4267-9a51-50e0d8325eaa","Type":"ContainerStarted","Data":"8b9bbb710bc44462ae6c623a7e30aab9f2ca50b16d3131eb393c5c2b7d42c33f"} Oct 01 16:09:33 crc kubenswrapper[4726]: I1001 16:09:33.181577 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-pz8pb" event={"ID":"55df3602-b3df-45f4-90b8-eddb1903bf32","Type":"ContainerStarted","Data":"358280423cb158770cd4e0607c02431c0828c5e09c562a3cd77fc7b54c5bfb02"} Oct 01 16:09:33 crc kubenswrapper[4726]: I1001 16:09:33.183809 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-jxd7m" event={"ID":"51706d0c-886c-4865-ba8b-ade37f39b84e","Type":"ContainerStarted","Data":"6e4f97a04f7876d73423d7e2ab33c4e33b74960b23560cf18c703c8ea6eccc2b"} Oct 01 16:09:33 crc kubenswrapper[4726]: I1001 16:09:33.184130 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-jxd7m" Oct 01 16:09:33 crc kubenswrapper[4726]: I1001 16:09:33.201225 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-jkfpv" podStartSLOduration=1.759078189 podStartE2EDuration="5.201197972s" podCreationTimestamp="2025-10-01 16:09:28 +0000 UTC" firstStartedPulling="2025-10-01 16:09:28.640403229 +0000 UTC m=+621.541955806" lastFinishedPulling="2025-10-01 16:09:32.082523012 +0000 UTC m=+624.984075589" observedRunningTime="2025-10-01 16:09:33.198251975 +0000 UTC m=+626.099804612" watchObservedRunningTime="2025-10-01 16:09:33.201197972 +0000 UTC m=+626.102750589" Oct 01 16:09:33 crc kubenswrapper[4726]: I1001 16:09:33.234604 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-jxd7m" podStartSLOduration=1.859508646 podStartE2EDuration="5.234580535s" podCreationTimestamp="2025-10-01 16:09:28 +0000 UTC" firstStartedPulling="2025-10-01 16:09:28.718246924 +0000 UTC m=+621.619799491" lastFinishedPulling="2025-10-01 16:09:32.093318803 +0000 UTC m=+624.994871380" observedRunningTime="2025-10-01 16:09:33.229911126 +0000 UTC m=+626.131463733" watchObservedRunningTime="2025-10-01 16:09:33.234580535 +0000 UTC m=+626.136133122" Oct 01 16:09:33 crc kubenswrapper[4726]: I1001 16:09:33.251773 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-pz8pb" podStartSLOduration=2.027452241 podStartE2EDuration="5.251744116s" podCreationTimestamp="2025-10-01 16:09:28 +0000 UTC" firstStartedPulling="2025-10-01 16:09:28.697131486 +0000 UTC m=+621.598684063" lastFinishedPulling="2025-10-01 16:09:31.921423351 +0000 UTC m=+624.822975938" observedRunningTime="2025-10-01 16:09:33.249892151 +0000 UTC m=+626.151444768" watchObservedRunningTime="2025-10-01 16:09:33.251744116 +0000 UTC m=+626.153296773" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.340362 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zp4jh"] Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.342087 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovn-controller" containerID="cri-o://55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a" gracePeriod=30 Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.342186 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="nbdb" containerID="cri-o://6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286" gracePeriod=30 Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.342316 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovn-acl-logging" containerID="cri-o://abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8" gracePeriod=30 Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.342303 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="northd" containerID="cri-o://76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413" gracePeriod=30 Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.342487 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="sbdb" containerID="cri-o://188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc" gracePeriod=30 Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.342280 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="kube-rbac-proxy-node" containerID="cri-o://cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788" gracePeriod=30 Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.342684 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714" gracePeriod=30 Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.395916 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" containerID="cri-o://6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8" gracePeriod=30 Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.449752 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-jxd7m" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.680788 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/3.log" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.683018 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovn-acl-logging/0.log" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.683585 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovn-controller/0.log" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.684012 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.748823 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bdt2r"] Oct 01 16:09:38 crc kubenswrapper[4726]: E1001 16:09:38.749146 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovn-acl-logging" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749172 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovn-acl-logging" Oct 01 16:09:38 crc kubenswrapper[4726]: E1001 16:09:38.749193 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovn-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749205 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovn-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: E1001 16:09:38.749226 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749237 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: E1001 16:09:38.749252 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="kubecfg-setup" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749263 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="kubecfg-setup" Oct 01 16:09:38 crc kubenswrapper[4726]: E1001 16:09:38.749280 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="kube-rbac-proxy-node" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749292 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="kube-rbac-proxy-node" Oct 01 16:09:38 crc kubenswrapper[4726]: E1001 16:09:38.749305 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="sbdb" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749318 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="sbdb" Oct 01 16:09:38 crc kubenswrapper[4726]: E1001 16:09:38.749336 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749349 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 16:09:38 crc kubenswrapper[4726]: E1001 16:09:38.749366 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="nbdb" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749379 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="nbdb" Oct 01 16:09:38 crc kubenswrapper[4726]: E1001 16:09:38.749392 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749403 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: E1001 16:09:38.749416 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749427 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: E1001 16:09:38.749444 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="northd" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749455 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="northd" Oct 01 16:09:38 crc kubenswrapper[4726]: E1001 16:09:38.749477 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749489 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749639 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749659 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovn-acl-logging" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749674 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749688 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="northd" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749699 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749712 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="nbdb" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749726 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovn-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749741 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="sbdb" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749757 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="kube-rbac-proxy-node" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749769 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749781 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 16:09:38 crc kubenswrapper[4726]: E1001 16:09:38.749950 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.749965 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.750144 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerName="ovnkube-controller" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.752884 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.761463 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-var-lib-openvswitch\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.761526 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-cni-netd\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.761584 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovnkube-config\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.761630 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-slash\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.761662 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.761701 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-kubelet\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.761733 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-openvswitch\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.761775 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovnkube-script-lib\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.761806 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-log-socket\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.761837 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znnhm\" (UniqueName: \"kubernetes.io/projected/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-kube-api-access-znnhm\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.761868 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-systemd\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.761904 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-ovn\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.761935 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-node-log\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.761970 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-env-overrides\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762004 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-cni-bin\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762042 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-run-ovn-kubernetes\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762124 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-systemd-units\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762163 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-run-netns\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762196 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovn-node-metrics-cert\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762228 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-etc-openvswitch\") pod \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\" (UID: \"48c2aeef-b0f7-41b1-8a03-519171bc0a6a\") " Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762380 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762403 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762468 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762496 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762511 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762534 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762545 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762602 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762646 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762662 4726 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762687 4726 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762691 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762709 4726 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762728 4726 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762746 4726 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762786 4726 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762727 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-node-log" (OuterVolumeSpecName: "node-log") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762802 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-slash" (OuterVolumeSpecName: "host-slash") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762782 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762836 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-log-socket" (OuterVolumeSpecName: "log-socket") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.762859 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.763041 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.763374 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.769791 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.770468 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-kube-api-access-znnhm" (OuterVolumeSpecName: "kube-api-access-znnhm") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "kube-api-access-znnhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.790099 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "48c2aeef-b0f7-41b1-8a03-519171bc0a6a" (UID: "48c2aeef-b0f7-41b1-8a03-519171bc0a6a"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863459 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-run-openvswitch\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863536 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-run-ovn-kubernetes\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863565 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-systemd-units\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863588 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-ovnkube-script-lib\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863611 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-slash\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863632 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-run-ovn\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863653 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-cni-bin\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863683 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-cni-netd\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863706 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-node-log\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863729 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-log-socket\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863760 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-ovnkube-config\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863782 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-run-systemd\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863800 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-run-netns\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863818 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-var-lib-openvswitch\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863839 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-env-overrides\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863865 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-ovn-node-metrics-cert\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863889 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-etc-openvswitch\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863909 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863931 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-kubelet\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863950 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgr57\" (UniqueName: \"kubernetes.io/projected/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-kube-api-access-tgr57\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.863990 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.864004 4726 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-log-socket\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.864015 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znnhm\" (UniqueName: \"kubernetes.io/projected/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-kube-api-access-znnhm\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.864026 4726 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.864037 4726 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.864075 4726 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-node-log\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.864086 4726 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.864096 4726 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.864107 4726 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.864117 4726 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.864128 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.864138 4726 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.864192 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.864230 4726 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48c2aeef-b0f7-41b1-8a03-519171bc0a6a-host-slash\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.964694 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-env-overrides\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.965116 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-ovn-node-metrics-cert\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.965383 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-etc-openvswitch\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.965553 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-etc-openvswitch\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.965626 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.965779 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.966147 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-env-overrides\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.966284 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-kubelet\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.966106 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-kubelet\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.966612 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgr57\" (UniqueName: \"kubernetes.io/projected/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-kube-api-access-tgr57\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.966821 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-run-openvswitch\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.967090 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-run-openvswitch\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.967115 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-run-ovn-kubernetes\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.967505 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-systemd-units\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.967734 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-ovnkube-script-lib\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.967949 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-slash\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.968197 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-run-ovn\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.968476 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-cni-bin\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.968745 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-cni-netd\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.968986 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-node-log\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.969670 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-log-socket\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.969835 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-log-socket\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.968552 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-cni-bin\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.968037 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-slash\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.968795 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-cni-netd\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.967306 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-run-ovn-kubernetes\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.969096 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-node-log\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.969235 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-ovnkube-script-lib\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.968257 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-run-ovn\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.967578 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-systemd-units\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.970529 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-ovn-node-metrics-cert\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.970805 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-ovnkube-config\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.971040 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-run-netns\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.971309 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-run-systemd\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.971515 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-var-lib-openvswitch\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.971703 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-run-systemd\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.971623 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-var-lib-openvswitch\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.971228 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-host-run-netns\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.972629 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-ovnkube-config\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:38 crc kubenswrapper[4726]: I1001 16:09:38.995279 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgr57\" (UniqueName: \"kubernetes.io/projected/cfc3a1e5-c3d7-49ca-8fa8-813aae28345f-kube-api-access-tgr57\") pod \"ovnkube-node-bdt2r\" (UID: \"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.068121 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:39 crc kubenswrapper[4726]: W1001 16:09:39.097890 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfc3a1e5_c3d7_49ca_8fa8_813aae28345f.slice/crio-daa64eacc2e921a2750f8f19a506539ca1f36f63262a6ab77cd27b3f0e2e5f8c WatchSource:0}: Error finding container daa64eacc2e921a2750f8f19a506539ca1f36f63262a6ab77cd27b3f0e2e5f8c: Status 404 returned error can't find the container with id daa64eacc2e921a2750f8f19a506539ca1f36f63262a6ab77cd27b3f0e2e5f8c Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.225205 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h8cql_e9257aec-e319-4261-ae4c-dec88468b680/kube-multus/2.log" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.226152 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h8cql_e9257aec-e319-4261-ae4c-dec88468b680/kube-multus/1.log" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.226237 4726 generic.go:334] "Generic (PLEG): container finished" podID="e9257aec-e319-4261-ae4c-dec88468b680" containerID="c061620e1547381b1e62f4d7068ac10ea23bd2fb250a3ce42f6f14cd2e8c727b" exitCode=2 Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.226304 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h8cql" event={"ID":"e9257aec-e319-4261-ae4c-dec88468b680","Type":"ContainerDied","Data":"c061620e1547381b1e62f4d7068ac10ea23bd2fb250a3ce42f6f14cd2e8c727b"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.226400 4726 scope.go:117] "RemoveContainer" containerID="e127584cf73c7f9ea20da97dce5f50fb86b8f9fde6ae383172834d425bb1e033" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.227277 4726 scope.go:117] "RemoveContainer" containerID="c061620e1547381b1e62f4d7068ac10ea23bd2fb250a3ce42f6f14cd2e8c727b" Oct 01 16:09:39 crc kubenswrapper[4726]: E1001 16:09:39.227880 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-h8cql_openshift-multus(e9257aec-e319-4261-ae4c-dec88468b680)\"" pod="openshift-multus/multus-h8cql" podUID="e9257aec-e319-4261-ae4c-dec88468b680" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.233572 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovnkube-controller/3.log" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.239108 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovn-acl-logging/0.log" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.240525 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zp4jh_48c2aeef-b0f7-41b1-8a03-519171bc0a6a/ovn-controller/0.log" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.241616 4726 generic.go:334] "Generic (PLEG): container finished" podID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerID="6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8" exitCode=0 Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.241664 4726 generic.go:334] "Generic (PLEG): container finished" podID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerID="188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc" exitCode=0 Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.241682 4726 generic.go:334] "Generic (PLEG): container finished" podID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerID="6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286" exitCode=0 Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.241702 4726 generic.go:334] "Generic (PLEG): container finished" podID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerID="76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413" exitCode=0 Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.241717 4726 generic.go:334] "Generic (PLEG): container finished" podID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerID="e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714" exitCode=0 Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.241730 4726 generic.go:334] "Generic (PLEG): container finished" podID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerID="cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788" exitCode=0 Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.241744 4726 generic.go:334] "Generic (PLEG): container finished" podID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerID="abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8" exitCode=143 Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.241758 4726 generic.go:334] "Generic (PLEG): container finished" podID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" containerID="55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a" exitCode=143 Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.241952 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242404 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerDied","Data":"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242454 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerDied","Data":"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242478 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerDied","Data":"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242502 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerDied","Data":"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242522 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerDied","Data":"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242541 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerDied","Data":"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242560 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242581 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242594 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242607 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242620 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242632 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242644 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242655 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242667 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242679 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242696 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerDied","Data":"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242715 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242730 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242745 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242757 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242768 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242779 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242792 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242804 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242816 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242827 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242842 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerDied","Data":"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242858 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242873 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242885 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242896 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242907 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242918 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242929 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242940 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242951 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242962 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242978 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zp4jh" event={"ID":"48c2aeef-b0f7-41b1-8a03-519171bc0a6a","Type":"ContainerDied","Data":"9053933dd57bfcbf4a57a5ab06bf1cb49e721929ae909665f0042374df9d604f"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.242996 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.243009 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.243021 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.243033 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.243045 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.243087 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.243100 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.243111 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.243124 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.243137 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.243715 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" event={"ID":"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f","Type":"ContainerStarted","Data":"daa64eacc2e921a2750f8f19a506539ca1f36f63262a6ab77cd27b3f0e2e5f8c"} Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.286695 4726 scope.go:117] "RemoveContainer" containerID="6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.316096 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zp4jh"] Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.318990 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zp4jh"] Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.378408 4726 scope.go:117] "RemoveContainer" containerID="04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.398490 4726 scope.go:117] "RemoveContainer" containerID="188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.410239 4726 scope.go:117] "RemoveContainer" containerID="6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.421233 4726 scope.go:117] "RemoveContainer" containerID="76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.432956 4726 scope.go:117] "RemoveContainer" containerID="e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.444378 4726 scope.go:117] "RemoveContainer" containerID="cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.460296 4726 scope.go:117] "RemoveContainer" containerID="abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.489256 4726 scope.go:117] "RemoveContainer" containerID="55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.500914 4726 scope.go:117] "RemoveContainer" containerID="b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.511392 4726 scope.go:117] "RemoveContainer" containerID="6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8" Oct 01 16:09:39 crc kubenswrapper[4726]: E1001 16:09:39.511879 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8\": container with ID starting with 6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8 not found: ID does not exist" containerID="6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.511918 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8"} err="failed to get container status \"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8\": rpc error: code = NotFound desc = could not find container \"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8\": container with ID starting with 6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.511943 4726 scope.go:117] "RemoveContainer" containerID="04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be" Oct 01 16:09:39 crc kubenswrapper[4726]: E1001 16:09:39.512440 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\": container with ID starting with 04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be not found: ID does not exist" containerID="04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.512480 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be"} err="failed to get container status \"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\": rpc error: code = NotFound desc = could not find container \"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\": container with ID starting with 04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.512493 4726 scope.go:117] "RemoveContainer" containerID="188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc" Oct 01 16:09:39 crc kubenswrapper[4726]: E1001 16:09:39.512780 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\": container with ID starting with 188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc not found: ID does not exist" containerID="188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.512849 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc"} err="failed to get container status \"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\": rpc error: code = NotFound desc = could not find container \"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\": container with ID starting with 188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.512873 4726 scope.go:117] "RemoveContainer" containerID="6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286" Oct 01 16:09:39 crc kubenswrapper[4726]: E1001 16:09:39.513235 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\": container with ID starting with 6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286 not found: ID does not exist" containerID="6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.513351 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286"} err="failed to get container status \"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\": rpc error: code = NotFound desc = could not find container \"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\": container with ID starting with 6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.513381 4726 scope.go:117] "RemoveContainer" containerID="76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413" Oct 01 16:09:39 crc kubenswrapper[4726]: E1001 16:09:39.513790 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\": container with ID starting with 76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413 not found: ID does not exist" containerID="76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.513831 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413"} err="failed to get container status \"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\": rpc error: code = NotFound desc = could not find container \"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\": container with ID starting with 76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.513864 4726 scope.go:117] "RemoveContainer" containerID="e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714" Oct 01 16:09:39 crc kubenswrapper[4726]: E1001 16:09:39.514230 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\": container with ID starting with e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714 not found: ID does not exist" containerID="e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.514256 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714"} err="failed to get container status \"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\": rpc error: code = NotFound desc = could not find container \"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\": container with ID starting with e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.514271 4726 scope.go:117] "RemoveContainer" containerID="cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788" Oct 01 16:09:39 crc kubenswrapper[4726]: E1001 16:09:39.514623 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\": container with ID starting with cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788 not found: ID does not exist" containerID="cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.514667 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788"} err="failed to get container status \"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\": rpc error: code = NotFound desc = could not find container \"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\": container with ID starting with cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.514695 4726 scope.go:117] "RemoveContainer" containerID="abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8" Oct 01 16:09:39 crc kubenswrapper[4726]: E1001 16:09:39.515000 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\": container with ID starting with abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8 not found: ID does not exist" containerID="abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.515026 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8"} err="failed to get container status \"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\": rpc error: code = NotFound desc = could not find container \"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\": container with ID starting with abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.515041 4726 scope.go:117] "RemoveContainer" containerID="55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a" Oct 01 16:09:39 crc kubenswrapper[4726]: E1001 16:09:39.515286 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\": container with ID starting with 55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a not found: ID does not exist" containerID="55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.515347 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a"} err="failed to get container status \"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\": rpc error: code = NotFound desc = could not find container \"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\": container with ID starting with 55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.515372 4726 scope.go:117] "RemoveContainer" containerID="b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a" Oct 01 16:09:39 crc kubenswrapper[4726]: E1001 16:09:39.515730 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\": container with ID starting with b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a not found: ID does not exist" containerID="b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.515751 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a"} err="failed to get container status \"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\": rpc error: code = NotFound desc = could not find container \"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\": container with ID starting with b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.515768 4726 scope.go:117] "RemoveContainer" containerID="6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.516240 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8"} err="failed to get container status \"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8\": rpc error: code = NotFound desc = could not find container \"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8\": container with ID starting with 6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.516277 4726 scope.go:117] "RemoveContainer" containerID="04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.516547 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be"} err="failed to get container status \"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\": rpc error: code = NotFound desc = could not find container \"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\": container with ID starting with 04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.516571 4726 scope.go:117] "RemoveContainer" containerID="188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.516847 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc"} err="failed to get container status \"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\": rpc error: code = NotFound desc = could not find container \"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\": container with ID starting with 188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.516867 4726 scope.go:117] "RemoveContainer" containerID="6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.517310 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286"} err="failed to get container status \"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\": rpc error: code = NotFound desc = could not find container \"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\": container with ID starting with 6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.517331 4726 scope.go:117] "RemoveContainer" containerID="76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.517701 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413"} err="failed to get container status \"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\": rpc error: code = NotFound desc = could not find container \"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\": container with ID starting with 76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.517719 4726 scope.go:117] "RemoveContainer" containerID="e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.517930 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714"} err="failed to get container status \"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\": rpc error: code = NotFound desc = could not find container \"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\": container with ID starting with e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.517949 4726 scope.go:117] "RemoveContainer" containerID="cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.518236 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788"} err="failed to get container status \"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\": rpc error: code = NotFound desc = could not find container \"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\": container with ID starting with cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.518280 4726 scope.go:117] "RemoveContainer" containerID="abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.518547 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8"} err="failed to get container status \"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\": rpc error: code = NotFound desc = could not find container \"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\": container with ID starting with abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.518563 4726 scope.go:117] "RemoveContainer" containerID="55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.518749 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a"} err="failed to get container status \"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\": rpc error: code = NotFound desc = could not find container \"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\": container with ID starting with 55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.518761 4726 scope.go:117] "RemoveContainer" containerID="b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.518994 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a"} err="failed to get container status \"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\": rpc error: code = NotFound desc = could not find container \"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\": container with ID starting with b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.519023 4726 scope.go:117] "RemoveContainer" containerID="6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.519318 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8"} err="failed to get container status \"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8\": rpc error: code = NotFound desc = could not find container \"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8\": container with ID starting with 6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.519339 4726 scope.go:117] "RemoveContainer" containerID="04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.519575 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be"} err="failed to get container status \"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\": rpc error: code = NotFound desc = could not find container \"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\": container with ID starting with 04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.519600 4726 scope.go:117] "RemoveContainer" containerID="188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.519839 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc"} err="failed to get container status \"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\": rpc error: code = NotFound desc = could not find container \"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\": container with ID starting with 188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.519858 4726 scope.go:117] "RemoveContainer" containerID="6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.520141 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286"} err="failed to get container status \"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\": rpc error: code = NotFound desc = could not find container \"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\": container with ID starting with 6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.520169 4726 scope.go:117] "RemoveContainer" containerID="76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.520423 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413"} err="failed to get container status \"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\": rpc error: code = NotFound desc = could not find container \"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\": container with ID starting with 76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.520460 4726 scope.go:117] "RemoveContainer" containerID="e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.527175 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714"} err="failed to get container status \"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\": rpc error: code = NotFound desc = could not find container \"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\": container with ID starting with e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.527220 4726 scope.go:117] "RemoveContainer" containerID="cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.527913 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788"} err="failed to get container status \"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\": rpc error: code = NotFound desc = could not find container \"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\": container with ID starting with cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.527934 4726 scope.go:117] "RemoveContainer" containerID="abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.528804 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8"} err="failed to get container status \"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\": rpc error: code = NotFound desc = could not find container \"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\": container with ID starting with abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.528839 4726 scope.go:117] "RemoveContainer" containerID="55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.529070 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a"} err="failed to get container status \"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\": rpc error: code = NotFound desc = could not find container \"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\": container with ID starting with 55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.529092 4726 scope.go:117] "RemoveContainer" containerID="b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.529382 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a"} err="failed to get container status \"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\": rpc error: code = NotFound desc = could not find container \"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\": container with ID starting with b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.529423 4726 scope.go:117] "RemoveContainer" containerID="6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.529725 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8"} err="failed to get container status \"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8\": rpc error: code = NotFound desc = could not find container \"6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8\": container with ID starting with 6b652e8e3b209a6a4b504ef1b62ab17a91ec9fdec86e12db7caa03a418bdd0b8 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.529755 4726 scope.go:117] "RemoveContainer" containerID="04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.530134 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be"} err="failed to get container status \"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\": rpc error: code = NotFound desc = could not find container \"04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be\": container with ID starting with 04ef7cea10b9e476d384d8c3c29af944b3d84f0431f4473b53b5d44de8ce05be not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.530165 4726 scope.go:117] "RemoveContainer" containerID="188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.530440 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc"} err="failed to get container status \"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\": rpc error: code = NotFound desc = could not find container \"188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc\": container with ID starting with 188fbcc63b3682692ade1c822e457dbfa36e22480ec2bd6c135a673fbe0e92cc not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.530461 4726 scope.go:117] "RemoveContainer" containerID="6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.530713 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286"} err="failed to get container status \"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\": rpc error: code = NotFound desc = could not find container \"6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286\": container with ID starting with 6226837dae67c9023ebdde737b789e3ff43c4d360c0d47f80ecade7bd29a8286 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.530737 4726 scope.go:117] "RemoveContainer" containerID="76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.530966 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413"} err="failed to get container status \"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\": rpc error: code = NotFound desc = could not find container \"76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413\": container with ID starting with 76e9a7c80aad9bd629f3a2c123c1e6a2d5fa7b77165343f892ef8e5f6d7b7413 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.531014 4726 scope.go:117] "RemoveContainer" containerID="e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.532024 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714"} err="failed to get container status \"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\": rpc error: code = NotFound desc = could not find container \"e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714\": container with ID starting with e1b86e5698101666bafa23f450c3e35f725e5c5905627b3ea8347f2d8e3ec714 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.532073 4726 scope.go:117] "RemoveContainer" containerID="cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.532297 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788"} err="failed to get container status \"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\": rpc error: code = NotFound desc = could not find container \"cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788\": container with ID starting with cff2272a73a281d04d2e62952b4621c9ad17d3511c242dc83f9b3bfaa5a7c788 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.532322 4726 scope.go:117] "RemoveContainer" containerID="abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.533684 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8"} err="failed to get container status \"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\": rpc error: code = NotFound desc = could not find container \"abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8\": container with ID starting with abf192d5af824c57f4d92d0f8a842fdd82db9b1952147dac798ca84a8acc3be8 not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.533706 4726 scope.go:117] "RemoveContainer" containerID="55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.534017 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a"} err="failed to get container status \"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\": rpc error: code = NotFound desc = could not find container \"55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a\": container with ID starting with 55e7125ac598b49cfc539df957c7dd1fa7caad171121e5dd495ca93d9d8fb36a not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.534041 4726 scope.go:117] "RemoveContainer" containerID="b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.535419 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a"} err="failed to get container status \"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\": rpc error: code = NotFound desc = could not find container \"b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a\": container with ID starting with b3109c3d808d9f1030fdbaa03f53adb3cd538ec48f74e2b8cfdd8b474cd72b8a not found: ID does not exist" Oct 01 16:09:39 crc kubenswrapper[4726]: I1001 16:09:39.819941 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48c2aeef-b0f7-41b1-8a03-519171bc0a6a" path="/var/lib/kubelet/pods/48c2aeef-b0f7-41b1-8a03-519171bc0a6a/volumes" Oct 01 16:09:40 crc kubenswrapper[4726]: I1001 16:09:40.255140 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h8cql_e9257aec-e319-4261-ae4c-dec88468b680/kube-multus/2.log" Oct 01 16:09:40 crc kubenswrapper[4726]: I1001 16:09:40.259811 4726 generic.go:334] "Generic (PLEG): container finished" podID="cfc3a1e5-c3d7-49ca-8fa8-813aae28345f" containerID="0ed605c7122c4a50ce9f341e4e4a00b47f93bd55bce20bda850591fd5c160f1a" exitCode=0 Oct 01 16:09:40 crc kubenswrapper[4726]: I1001 16:09:40.259961 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" event={"ID":"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f","Type":"ContainerDied","Data":"0ed605c7122c4a50ce9f341e4e4a00b47f93bd55bce20bda850591fd5c160f1a"} Oct 01 16:09:41 crc kubenswrapper[4726]: I1001 16:09:41.268426 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" event={"ID":"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f","Type":"ContainerStarted","Data":"3bb3c45f8b97c08178cefd904c701eea738bfffd0afecfb1721bb056b6cb351b"} Oct 01 16:09:41 crc kubenswrapper[4726]: I1001 16:09:41.268930 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" event={"ID":"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f","Type":"ContainerStarted","Data":"eaff419f8e57d4e7269cd4da61784922f7e3d3cca146db99dee1f382c85b1211"} Oct 01 16:09:41 crc kubenswrapper[4726]: I1001 16:09:41.268941 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" event={"ID":"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f","Type":"ContainerStarted","Data":"c148372db2d80f512d61545947fe0afb3b01af237220b34ce33028dbcf93a857"} Oct 01 16:09:41 crc kubenswrapper[4726]: I1001 16:09:41.268951 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" event={"ID":"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f","Type":"ContainerStarted","Data":"9910335fb0269e99eac14b116c9d0952138160563baf5d19ad1b0908eee09466"} Oct 01 16:09:41 crc kubenswrapper[4726]: I1001 16:09:41.268961 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" event={"ID":"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f","Type":"ContainerStarted","Data":"52bf027fa8b07ce48a794d094b3f950af3e445fcacaac03e4a331d6383a8281b"} Oct 01 16:09:42 crc kubenswrapper[4726]: I1001 16:09:42.280019 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" event={"ID":"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f","Type":"ContainerStarted","Data":"4f66757e94cb9eeae7861be2d516355b1b7c6eb84c8fa6c67da00086cba62451"} Oct 01 16:09:44 crc kubenswrapper[4726]: I1001 16:09:44.296911 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" event={"ID":"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f","Type":"ContainerStarted","Data":"fc658bb90f7c4c497082a574802b8240281d1fd4b7a56437150997e432b8a3b8"} Oct 01 16:09:46 crc kubenswrapper[4726]: I1001 16:09:46.313695 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" event={"ID":"cfc3a1e5-c3d7-49ca-8fa8-813aae28345f","Type":"ContainerStarted","Data":"da3c4e3673197d1546cce705a0cb100d888928b24357557f67fa8594afb5ba6b"} Oct 01 16:09:46 crc kubenswrapper[4726]: I1001 16:09:46.314124 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:46 crc kubenswrapper[4726]: I1001 16:09:46.340461 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" podStartSLOduration=8.340442772 podStartE2EDuration="8.340442772s" podCreationTimestamp="2025-10-01 16:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:09:46.337171895 +0000 UTC m=+639.238724492" watchObservedRunningTime="2025-10-01 16:09:46.340442772 +0000 UTC m=+639.241995369" Oct 01 16:09:46 crc kubenswrapper[4726]: I1001 16:09:46.355216 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:47 crc kubenswrapper[4726]: I1001 16:09:47.320201 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:47 crc kubenswrapper[4726]: I1001 16:09:47.320281 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:47 crc kubenswrapper[4726]: I1001 16:09:47.360158 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:09:51 crc kubenswrapper[4726]: I1001 16:09:51.808294 4726 scope.go:117] "RemoveContainer" containerID="c061620e1547381b1e62f4d7068ac10ea23bd2fb250a3ce42f6f14cd2e8c727b" Oct 01 16:09:51 crc kubenswrapper[4726]: E1001 16:09:51.808929 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-h8cql_openshift-multus(e9257aec-e319-4261-ae4c-dec88468b680)\"" pod="openshift-multus/multus-h8cql" podUID="e9257aec-e319-4261-ae4c-dec88468b680" Oct 01 16:10:05 crc kubenswrapper[4726]: I1001 16:10:05.808028 4726 scope.go:117] "RemoveContainer" containerID="c061620e1547381b1e62f4d7068ac10ea23bd2fb250a3ce42f6f14cd2e8c727b" Oct 01 16:10:06 crc kubenswrapper[4726]: I1001 16:10:06.445093 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h8cql_e9257aec-e319-4261-ae4c-dec88468b680/kube-multus/2.log" Oct 01 16:10:06 crc kubenswrapper[4726]: I1001 16:10:06.445302 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h8cql" event={"ID":"e9257aec-e319-4261-ae4c-dec88468b680","Type":"ContainerStarted","Data":"4c05f4825545727af4d940bd11e835e2578f8e5eeb1f8d04d1b646fb1e7f5a5a"} Oct 01 16:10:09 crc kubenswrapper[4726]: I1001 16:10:09.097735 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bdt2r" Oct 01 16:10:18 crc kubenswrapper[4726]: I1001 16:10:18.842978 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns"] Oct 01 16:10:18 crc kubenswrapper[4726]: I1001 16:10:18.844529 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" Oct 01 16:10:18 crc kubenswrapper[4726]: I1001 16:10:18.847445 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 01 16:10:18 crc kubenswrapper[4726]: I1001 16:10:18.857354 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns"] Oct 01 16:10:18 crc kubenswrapper[4726]: I1001 16:10:18.991463 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns\" (UID: \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" Oct 01 16:10:18 crc kubenswrapper[4726]: I1001 16:10:18.992272 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns\" (UID: \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" Oct 01 16:10:18 crc kubenswrapper[4726]: I1001 16:10:18.992413 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qpqp\" (UniqueName: \"kubernetes.io/projected/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-kube-api-access-6qpqp\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns\" (UID: \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" Oct 01 16:10:19 crc kubenswrapper[4726]: I1001 16:10:19.093972 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns\" (UID: \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" Oct 01 16:10:19 crc kubenswrapper[4726]: I1001 16:10:19.094110 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns\" (UID: \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" Oct 01 16:10:19 crc kubenswrapper[4726]: I1001 16:10:19.094177 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qpqp\" (UniqueName: \"kubernetes.io/projected/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-kube-api-access-6qpqp\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns\" (UID: \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" Oct 01 16:10:19 crc kubenswrapper[4726]: I1001 16:10:19.094635 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns\" (UID: \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" Oct 01 16:10:19 crc kubenswrapper[4726]: I1001 16:10:19.095431 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns\" (UID: \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" Oct 01 16:10:19 crc kubenswrapper[4726]: I1001 16:10:19.130834 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qpqp\" (UniqueName: \"kubernetes.io/projected/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-kube-api-access-6qpqp\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns\" (UID: \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" Oct 01 16:10:19 crc kubenswrapper[4726]: I1001 16:10:19.165002 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" Oct 01 16:10:19 crc kubenswrapper[4726]: I1001 16:10:19.453708 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns"] Oct 01 16:10:19 crc kubenswrapper[4726]: I1001 16:10:19.524246 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" event={"ID":"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538","Type":"ContainerStarted","Data":"1c0430687e1d80ea8158ec0813ab041cb29885a17f718eb02fff9e3fc19c0de0"} Oct 01 16:10:20 crc kubenswrapper[4726]: I1001 16:10:20.534413 4726 generic.go:334] "Generic (PLEG): container finished" podID="f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" containerID="3b9144141cc84ecc4fc3bf1474ddc2bd5214807a8328ada4c230e53d5d38e11f" exitCode=0 Oct 01 16:10:20 crc kubenswrapper[4726]: I1001 16:10:20.534505 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" event={"ID":"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538","Type":"ContainerDied","Data":"3b9144141cc84ecc4fc3bf1474ddc2bd5214807a8328ada4c230e53d5d38e11f"} Oct 01 16:10:21 crc kubenswrapper[4726]: E1001 16:10:21.870169 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = reading blob sha256:7e15005ceea3703f0dbdcf6553dc41f9b15f72cf338bf44f5033a30b92fc41bf: Digest did not match, expected sha256:7e15005ceea3703f0dbdcf6553dc41f9b15f72cf338bf44f5033a30b92fc41bf, got sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" image="registry.redhat.io/openshift4/kubernetes-nmstate-operator-bundle@sha256:5b61484c6a2bbc377435093bd9abb31df980ece470325cae92f3106963b6482c" Oct 01 16:10:21 crc kubenswrapper[4726]: E1001 16:10:21.870393 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:pull,Image:registry.redhat.io/openshift4/kubernetes-nmstate-operator-bundle@sha256:5b61484c6a2bbc377435093bd9abb31df980ece470325cae92f3106963b6482c,Command:[/util/cpb /bundle],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:bundle,ReadOnly:false,MountPath:/bundle,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:util,ReadOnly:false,MountPath:/util,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6qpqp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod 9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_openshift-marketplace(f2e6c6e1-9224-4c26-adcf-07fdf1ea4538): ErrImagePull: reading blob sha256:7e15005ceea3703f0dbdcf6553dc41f9b15f72cf338bf44f5033a30b92fc41bf: Digest did not match, expected sha256:7e15005ceea3703f0dbdcf6553dc41f9b15f72cf338bf44f5033a30b92fc41bf, got sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" logger="UnhandledError" Oct 01 16:10:21 crc kubenswrapper[4726]: E1001 16:10:21.872539 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"pull\" with ErrImagePull: \"reading blob sha256:7e15005ceea3703f0dbdcf6553dc41f9b15f72cf338bf44f5033a30b92fc41bf: Digest did not match, expected sha256:7e15005ceea3703f0dbdcf6553dc41f9b15f72cf338bf44f5033a30b92fc41bf, got sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855\"" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" podUID="f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" Oct 01 16:10:22 crc kubenswrapper[4726]: E1001 16:10:22.547464 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"pull\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/kubernetes-nmstate-operator-bundle@sha256:5b61484c6a2bbc377435093bd9abb31df980ece470325cae92f3106963b6482c\\\"\"" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" podUID="f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" Oct 01 16:10:36 crc kubenswrapper[4726]: I1001 16:10:36.627727 4726 generic.go:334] "Generic (PLEG): container finished" podID="f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" containerID="39b03fca5f021ba970fb45bfb3ae49c5eeca8a6b604b06a88ee23e3fe557221b" exitCode=0 Oct 01 16:10:36 crc kubenswrapper[4726]: I1001 16:10:36.627894 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" event={"ID":"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538","Type":"ContainerDied","Data":"39b03fca5f021ba970fb45bfb3ae49c5eeca8a6b604b06a88ee23e3fe557221b"} Oct 01 16:10:37 crc kubenswrapper[4726]: I1001 16:10:37.636572 4726 generic.go:334] "Generic (PLEG): container finished" podID="f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" containerID="3de3fc2fc9623455c4a96cdea6bd5d28b859470dbff4bbcd3b277ee847fbc2ce" exitCode=0 Oct 01 16:10:37 crc kubenswrapper[4726]: I1001 16:10:37.636651 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" event={"ID":"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538","Type":"ContainerDied","Data":"3de3fc2fc9623455c4a96cdea6bd5d28b859470dbff4bbcd3b277ee847fbc2ce"} Oct 01 16:10:38 crc kubenswrapper[4726]: I1001 16:10:38.876124 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" Oct 01 16:10:38 crc kubenswrapper[4726]: I1001 16:10:38.985990 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-bundle\") pod \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\" (UID: \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\") " Oct 01 16:10:38 crc kubenswrapper[4726]: I1001 16:10:38.986112 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qpqp\" (UniqueName: \"kubernetes.io/projected/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-kube-api-access-6qpqp\") pod \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\" (UID: \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\") " Oct 01 16:10:38 crc kubenswrapper[4726]: I1001 16:10:38.986229 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-util\") pod \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\" (UID: \"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538\") " Oct 01 16:10:38 crc kubenswrapper[4726]: I1001 16:10:38.986686 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-bundle" (OuterVolumeSpecName: "bundle") pod "f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" (UID: "f2e6c6e1-9224-4c26-adcf-07fdf1ea4538"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:10:38 crc kubenswrapper[4726]: I1001 16:10:38.992216 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-kube-api-access-6qpqp" (OuterVolumeSpecName: "kube-api-access-6qpqp") pod "f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" (UID: "f2e6c6e1-9224-4c26-adcf-07fdf1ea4538"). InnerVolumeSpecName "kube-api-access-6qpqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:10:39 crc kubenswrapper[4726]: I1001 16:10:39.008439 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-util" (OuterVolumeSpecName: "util") pod "f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" (UID: "f2e6c6e1-9224-4c26-adcf-07fdf1ea4538"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:10:39 crc kubenswrapper[4726]: I1001 16:10:39.088028 4726 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-util\") on node \"crc\" DevicePath \"\"" Oct 01 16:10:39 crc kubenswrapper[4726]: I1001 16:10:39.088114 4726 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:10:39 crc kubenswrapper[4726]: I1001 16:10:39.088146 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qpqp\" (UniqueName: \"kubernetes.io/projected/f2e6c6e1-9224-4c26-adcf-07fdf1ea4538-kube-api-access-6qpqp\") on node \"crc\" DevicePath \"\"" Oct 01 16:10:39 crc kubenswrapper[4726]: I1001 16:10:39.653589 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" Oct 01 16:10:39 crc kubenswrapper[4726]: I1001 16:10:39.653555 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns" event={"ID":"f2e6c6e1-9224-4c26-adcf-07fdf1ea4538","Type":"ContainerDied","Data":"1c0430687e1d80ea8158ec0813ab041cb29885a17f718eb02fff9e3fc19c0de0"} Oct 01 16:10:39 crc kubenswrapper[4726]: I1001 16:10:39.654906 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c0430687e1d80ea8158ec0813ab041cb29885a17f718eb02fff9e3fc19c0de0" Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.349523 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8"] Oct 01 16:10:45 crc kubenswrapper[4726]: E1001 16:10:45.350147 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" containerName="extract" Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.350159 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" containerName="extract" Oct 01 16:10:45 crc kubenswrapper[4726]: E1001 16:10:45.350170 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" containerName="pull" Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.350176 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" containerName="pull" Oct 01 16:10:45 crc kubenswrapper[4726]: E1001 16:10:45.350184 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" containerName="util" Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.350191 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" containerName="util" Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.350302 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2e6c6e1-9224-4c26-adcf-07fdf1ea4538" containerName="extract" Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.350688 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8" Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.352961 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.354826 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-25xbl" Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.355264 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.367921 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8"] Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.472575 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtq2t\" (UniqueName: \"kubernetes.io/projected/5de9f82a-fe6f-47a8-be2a-0a513407b081-kube-api-access-rtq2t\") pod \"nmstate-operator-5d6f6cfd66-fk9n8\" (UID: \"5de9f82a-fe6f-47a8-be2a-0a513407b081\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8" Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.574296 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtq2t\" (UniqueName: \"kubernetes.io/projected/5de9f82a-fe6f-47a8-be2a-0a513407b081-kube-api-access-rtq2t\") pod \"nmstate-operator-5d6f6cfd66-fk9n8\" (UID: \"5de9f82a-fe6f-47a8-be2a-0a513407b081\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8" Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.608407 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtq2t\" (UniqueName: \"kubernetes.io/projected/5de9f82a-fe6f-47a8-be2a-0a513407b081-kube-api-access-rtq2t\") pod \"nmstate-operator-5d6f6cfd66-fk9n8\" (UID: \"5de9f82a-fe6f-47a8-be2a-0a513407b081\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8" Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.666714 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8" Oct 01 16:10:45 crc kubenswrapper[4726]: I1001 16:10:45.936626 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8"] Oct 01 16:10:45 crc kubenswrapper[4726]: W1001 16:10:45.942036 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5de9f82a_fe6f_47a8_be2a_0a513407b081.slice/crio-2f631e978f289cf0f0b69526f985aef51d8dbd2f909c18a06299a035e892f63f WatchSource:0}: Error finding container 2f631e978f289cf0f0b69526f985aef51d8dbd2f909c18a06299a035e892f63f: Status 404 returned error can't find the container with id 2f631e978f289cf0f0b69526f985aef51d8dbd2f909c18a06299a035e892f63f Oct 01 16:10:46 crc kubenswrapper[4726]: I1001 16:10:46.699378 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8" event={"ID":"5de9f82a-fe6f-47a8-be2a-0a513407b081","Type":"ContainerStarted","Data":"2f631e978f289cf0f0b69526f985aef51d8dbd2f909c18a06299a035e892f63f"} Oct 01 16:10:48 crc kubenswrapper[4726]: I1001 16:10:48.714521 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8" event={"ID":"5de9f82a-fe6f-47a8-be2a-0a513407b081","Type":"ContainerStarted","Data":"f3626a99bbcbca80cdd89a6cbd66f2c1b6cadb8839c400ebf05eef48c1350422"} Oct 01 16:10:48 crc kubenswrapper[4726]: I1001 16:10:48.742661 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8" podStartSLOduration=1.6361500169999998 podStartE2EDuration="3.742637791s" podCreationTimestamp="2025-10-01 16:10:45 +0000 UTC" firstStartedPulling="2025-10-01 16:10:45.944637321 +0000 UTC m=+698.846189898" lastFinishedPulling="2025-10-01 16:10:48.051125095 +0000 UTC m=+700.952677672" observedRunningTime="2025-10-01 16:10:48.737450658 +0000 UTC m=+701.639003315" watchObservedRunningTime="2025-10-01 16:10:48.742637791 +0000 UTC m=+701.644190398" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.632219 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz"] Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.633247 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.635130 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-fhj2m" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.639502 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz"] Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.646150 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-zftsg"] Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.647142 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.684760 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.692895 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-zftsg"] Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.718081 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-bg7hk"] Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.718663 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.727613 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzhfn\" (UniqueName: \"kubernetes.io/projected/f179fe18-0a34-43df-bb39-dab69a9ee29b-kube-api-access-fzhfn\") pod \"nmstate-metrics-58fcddf996-mdcjz\" (UID: \"f179fe18-0a34-43df-bb39-dab69a9ee29b\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.727767 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bkh9\" (UniqueName: \"kubernetes.io/projected/4c168234-66ba-492c-986c-835301bd0275-kube-api-access-4bkh9\") pod \"nmstate-webhook-6d689559c5-zftsg\" (UID: \"4c168234-66ba-492c-986c-835301bd0275\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.728475 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4c168234-66ba-492c-986c-835301bd0275-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-zftsg\" (UID: \"4c168234-66ba-492c-986c-835301bd0275\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.800029 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh"] Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.800838 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.802609 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.803149 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.807441 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh"] Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.809553 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-rrwm6" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.829835 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4c168234-66ba-492c-986c-835301bd0275-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-zftsg\" (UID: \"4c168234-66ba-492c-986c-835301bd0275\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.829885 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-ovs-socket\") pod \"nmstate-handler-bg7hk\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.829936 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-nmstate-lock\") pod \"nmstate-handler-bg7hk\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.829958 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-275r5\" (UniqueName: \"kubernetes.io/projected/98229bdb-bd95-46a7-9b03-04698c7f2475-kube-api-access-275r5\") pod \"nmstate-handler-bg7hk\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.829978 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzhfn\" (UniqueName: \"kubernetes.io/projected/f179fe18-0a34-43df-bb39-dab69a9ee29b-kube-api-access-fzhfn\") pod \"nmstate-metrics-58fcddf996-mdcjz\" (UID: \"f179fe18-0a34-43df-bb39-dab69a9ee29b\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.829998 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bkh9\" (UniqueName: \"kubernetes.io/projected/4c168234-66ba-492c-986c-835301bd0275-kube-api-access-4bkh9\") pod \"nmstate-webhook-6d689559c5-zftsg\" (UID: \"4c168234-66ba-492c-986c-835301bd0275\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.830154 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-dbus-socket\") pod \"nmstate-handler-bg7hk\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:49 crc kubenswrapper[4726]: E1001 16:10:49.830042 4726 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 01 16:10:49 crc kubenswrapper[4726]: E1001 16:10:49.830379 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c168234-66ba-492c-986c-835301bd0275-tls-key-pair podName:4c168234-66ba-492c-986c-835301bd0275 nodeName:}" failed. No retries permitted until 2025-10-01 16:10:50.33036165 +0000 UTC m=+703.231914227 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/4c168234-66ba-492c-986c-835301bd0275-tls-key-pair") pod "nmstate-webhook-6d689559c5-zftsg" (UID: "4c168234-66ba-492c-986c-835301bd0275") : secret "openshift-nmstate-webhook" not found Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.846854 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzhfn\" (UniqueName: \"kubernetes.io/projected/f179fe18-0a34-43df-bb39-dab69a9ee29b-kube-api-access-fzhfn\") pod \"nmstate-metrics-58fcddf996-mdcjz\" (UID: \"f179fe18-0a34-43df-bb39-dab69a9ee29b\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.847085 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bkh9\" (UniqueName: \"kubernetes.io/projected/4c168234-66ba-492c-986c-835301bd0275-kube-api-access-4bkh9\") pod \"nmstate-webhook-6d689559c5-zftsg\" (UID: \"4c168234-66ba-492c-986c-835301bd0275\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.930798 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c725f6ed-2c36-43dc-86f4-3a535030832b-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-xffxh\" (UID: \"c725f6ed-2c36-43dc-86f4-3a535030832b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.930918 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-dbus-socket\") pod \"nmstate-handler-bg7hk\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.931175 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-ovs-socket\") pod \"nmstate-handler-bg7hk\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.931249 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c725f6ed-2c36-43dc-86f4-3a535030832b-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-xffxh\" (UID: \"c725f6ed-2c36-43dc-86f4-3a535030832b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.931273 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-dbus-socket\") pod \"nmstate-handler-bg7hk\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.931321 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-ovs-socket\") pod \"nmstate-handler-bg7hk\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.931370 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-nmstate-lock\") pod \"nmstate-handler-bg7hk\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.931466 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-nmstate-lock\") pod \"nmstate-handler-bg7hk\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.931519 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-275r5\" (UniqueName: \"kubernetes.io/projected/98229bdb-bd95-46a7-9b03-04698c7f2475-kube-api-access-275r5\") pod \"nmstate-handler-bg7hk\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.931635 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9rcr\" (UniqueName: \"kubernetes.io/projected/c725f6ed-2c36-43dc-86f4-3a535030832b-kube-api-access-d9rcr\") pod \"nmstate-console-plugin-864bb6dfb5-xffxh\" (UID: \"c725f6ed-2c36-43dc-86f4-3a535030832b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.948682 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-275r5\" (UniqueName: \"kubernetes.io/projected/98229bdb-bd95-46a7-9b03-04698c7f2475-kube-api-access-275r5\") pod \"nmstate-handler-bg7hk\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.988179 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5f4d7b984b-cpbsb"] Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.989037 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:49 crc kubenswrapper[4726]: I1001 16:10:49.989288 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.012734 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5f4d7b984b-cpbsb"] Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.035017 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-console-config\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.035094 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-console-serving-cert\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.035132 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-oauth-serving-cert\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.035171 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c725f6ed-2c36-43dc-86f4-3a535030832b-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-xffxh\" (UID: \"c725f6ed-2c36-43dc-86f4-3a535030832b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.035213 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-service-ca\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.035243 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-console-oauth-config\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.035272 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bx6l\" (UniqueName: \"kubernetes.io/projected/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-kube-api-access-8bx6l\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.035329 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9rcr\" (UniqueName: \"kubernetes.io/projected/c725f6ed-2c36-43dc-86f4-3a535030832b-kube-api-access-d9rcr\") pod \"nmstate-console-plugin-864bb6dfb5-xffxh\" (UID: \"c725f6ed-2c36-43dc-86f4-3a535030832b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.035369 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c725f6ed-2c36-43dc-86f4-3a535030832b-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-xffxh\" (UID: \"c725f6ed-2c36-43dc-86f4-3a535030832b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.035415 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-trusted-ca-bundle\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.036882 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c725f6ed-2c36-43dc-86f4-3a535030832b-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-xffxh\" (UID: \"c725f6ed-2c36-43dc-86f4-3a535030832b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.043351 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.048219 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c725f6ed-2c36-43dc-86f4-3a535030832b-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-xffxh\" (UID: \"c725f6ed-2c36-43dc-86f4-3a535030832b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.059317 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9rcr\" (UniqueName: \"kubernetes.io/projected/c725f6ed-2c36-43dc-86f4-3a535030832b-kube-api-access-d9rcr\") pod \"nmstate-console-plugin-864bb6dfb5-xffxh\" (UID: \"c725f6ed-2c36-43dc-86f4-3a535030832b\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.114122 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.136931 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-trusted-ca-bundle\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.137003 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-console-config\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.137029 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-console-serving-cert\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.137070 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-oauth-serving-cert\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.137106 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-service-ca\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.137126 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-console-oauth-config\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.137176 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bx6l\" (UniqueName: \"kubernetes.io/projected/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-kube-api-access-8bx6l\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.138433 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-oauth-serving-cert\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.138629 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-trusted-ca-bundle\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.139150 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-service-ca\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.139313 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-console-config\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.141960 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-console-oauth-config\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.142309 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-console-serving-cert\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.161118 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bx6l\" (UniqueName: \"kubernetes.io/projected/c5073622-b2b8-4efe-8b8c-9d3a0aa0506b-kube-api-access-8bx6l\") pod \"console-5f4d7b984b-cpbsb\" (UID: \"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b\") " pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.261637 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz"] Oct 01 16:10:50 crc kubenswrapper[4726]: W1001 16:10:50.270319 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf179fe18_0a34_43df_bb39_dab69a9ee29b.slice/crio-e0898ee0e6c932ae4b7ae6523d8d2c16d1455fbe37b5faa1de588406ec3c4159 WatchSource:0}: Error finding container e0898ee0e6c932ae4b7ae6523d8d2c16d1455fbe37b5faa1de588406ec3c4159: Status 404 returned error can't find the container with id e0898ee0e6c932ae4b7ae6523d8d2c16d1455fbe37b5faa1de588406ec3c4159 Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.331838 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh"] Oct 01 16:10:50 crc kubenswrapper[4726]: W1001 16:10:50.338510 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc725f6ed_2c36_43dc_86f4_3a535030832b.slice/crio-3140a050a2ca6affa03455f3cde5dcfbf5d065bce40d938018a23712119be6cf WatchSource:0}: Error finding container 3140a050a2ca6affa03455f3cde5dcfbf5d065bce40d938018a23712119be6cf: Status 404 returned error can't find the container with id 3140a050a2ca6affa03455f3cde5dcfbf5d065bce40d938018a23712119be6cf Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.340009 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.341384 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4c168234-66ba-492c-986c-835301bd0275-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-zftsg\" (UID: \"4c168234-66ba-492c-986c-835301bd0275\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.344545 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4c168234-66ba-492c-986c-835301bd0275-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-zftsg\" (UID: \"4c168234-66ba-492c-986c-835301bd0275\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.524074 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5f4d7b984b-cpbsb"] Oct 01 16:10:50 crc kubenswrapper[4726]: W1001 16:10:50.528454 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5073622_b2b8_4efe_8b8c_9d3a0aa0506b.slice/crio-777bc862aaa1a6b033a76488bce694e837326844c82c6d396be2534bcc370530 WatchSource:0}: Error finding container 777bc862aaa1a6b033a76488bce694e837326844c82c6d396be2534bcc370530: Status 404 returned error can't find the container with id 777bc862aaa1a6b033a76488bce694e837326844c82c6d396be2534bcc370530 Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.612702 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.730488 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-bg7hk" event={"ID":"98229bdb-bd95-46a7-9b03-04698c7f2475","Type":"ContainerStarted","Data":"7dab191f6b1d9e19fd289b3c01888b708ab9e6422c6dc51c08dad7acdac612b4"} Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.732003 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5f4d7b984b-cpbsb" event={"ID":"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b","Type":"ContainerStarted","Data":"cf9bca55e03163ec0e91d728c6862653f7482e9f95531f183265efab7f0146b7"} Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.732069 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5f4d7b984b-cpbsb" event={"ID":"c5073622-b2b8-4efe-8b8c-9d3a0aa0506b","Type":"ContainerStarted","Data":"777bc862aaa1a6b033a76488bce694e837326844c82c6d396be2534bcc370530"} Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.733991 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" event={"ID":"c725f6ed-2c36-43dc-86f4-3a535030832b","Type":"ContainerStarted","Data":"3140a050a2ca6affa03455f3cde5dcfbf5d065bce40d938018a23712119be6cf"} Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.736430 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" event={"ID":"f179fe18-0a34-43df-bb39-dab69a9ee29b","Type":"ContainerStarted","Data":"e0898ee0e6c932ae4b7ae6523d8d2c16d1455fbe37b5faa1de588406ec3c4159"} Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.751924 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5f4d7b984b-cpbsb" podStartSLOduration=1.75191 podStartE2EDuration="1.75191s" podCreationTimestamp="2025-10-01 16:10:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:10:50.749391586 +0000 UTC m=+703.650944163" watchObservedRunningTime="2025-10-01 16:10:50.75191 +0000 UTC m=+703.653462577" Oct 01 16:10:50 crc kubenswrapper[4726]: I1001 16:10:50.856076 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-zftsg"] Oct 01 16:10:50 crc kubenswrapper[4726]: W1001 16:10:50.859627 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c168234_66ba_492c_986c_835301bd0275.slice/crio-95d3ce4a10223541ca661d7d9592c3f3f9f461ca4bf4d06769171ba50bd08718 WatchSource:0}: Error finding container 95d3ce4a10223541ca661d7d9592c3f3f9f461ca4bf4d06769171ba50bd08718: Status 404 returned error can't find the container with id 95d3ce4a10223541ca661d7d9592c3f3f9f461ca4bf4d06769171ba50bd08718 Oct 01 16:10:51 crc kubenswrapper[4726]: I1001 16:10:51.745137 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" event={"ID":"4c168234-66ba-492c-986c-835301bd0275","Type":"ContainerStarted","Data":"95d3ce4a10223541ca661d7d9592c3f3f9f461ca4bf4d06769171ba50bd08718"} Oct 01 16:10:53 crc kubenswrapper[4726]: I1001 16:10:53.761252 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" event={"ID":"c725f6ed-2c36-43dc-86f4-3a535030832b","Type":"ContainerStarted","Data":"9702f03a2d467ad4bbc71677132d4711df89659823db12868d1296226c3c611f"} Oct 01 16:10:53 crc kubenswrapper[4726]: I1001 16:10:53.763190 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" event={"ID":"f179fe18-0a34-43df-bb39-dab69a9ee29b","Type":"ContainerStarted","Data":"15fff174bbe0d667779152131358e8d5810709cbfaa917040c43a8ad63a90369"} Oct 01 16:10:53 crc kubenswrapper[4726]: I1001 16:10:53.765359 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-bg7hk" event={"ID":"98229bdb-bd95-46a7-9b03-04698c7f2475","Type":"ContainerStarted","Data":"09ccffad3d0057482a969cd85fd8ae0902881bb7552b904cc6d0aa11df4a7da8"} Oct 01 16:10:53 crc kubenswrapper[4726]: I1001 16:10:53.765459 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:10:53 crc kubenswrapper[4726]: I1001 16:10:53.767479 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" event={"ID":"4c168234-66ba-492c-986c-835301bd0275","Type":"ContainerStarted","Data":"df28c3773d7ed079f09aceea751eca32bac74851679679cfa0d36e0a15abef2e"} Oct 01 16:10:53 crc kubenswrapper[4726]: I1001 16:10:53.767607 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" Oct 01 16:10:53 crc kubenswrapper[4726]: I1001 16:10:53.796852 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" podStartSLOduration=1.918316516 podStartE2EDuration="4.796795378s" podCreationTimestamp="2025-10-01 16:10:49 +0000 UTC" firstStartedPulling="2025-10-01 16:10:50.341473908 +0000 UTC m=+703.243026485" lastFinishedPulling="2025-10-01 16:10:53.21995276 +0000 UTC m=+706.121505347" observedRunningTime="2025-10-01 16:10:53.783643832 +0000 UTC m=+706.685196479" watchObservedRunningTime="2025-10-01 16:10:53.796795378 +0000 UTC m=+706.698348025" Oct 01 16:10:53 crc kubenswrapper[4726]: I1001 16:10:53.820016 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-bg7hk" podStartSLOduration=1.661292869 podStartE2EDuration="4.819983119s" podCreationTimestamp="2025-10-01 16:10:49 +0000 UTC" firstStartedPulling="2025-10-01 16:10:50.064101193 +0000 UTC m=+702.965653780" lastFinishedPulling="2025-10-01 16:10:53.222791453 +0000 UTC m=+706.124344030" observedRunningTime="2025-10-01 16:10:53.803135654 +0000 UTC m=+706.704688241" watchObservedRunningTime="2025-10-01 16:10:53.819983119 +0000 UTC m=+706.721535736" Oct 01 16:10:53 crc kubenswrapper[4726]: I1001 16:10:53.828933 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" podStartSLOduration=2.448639249 podStartE2EDuration="4.828911741s" podCreationTimestamp="2025-10-01 16:10:49 +0000 UTC" firstStartedPulling="2025-10-01 16:10:50.86224558 +0000 UTC m=+703.763798177" lastFinishedPulling="2025-10-01 16:10:53.242518082 +0000 UTC m=+706.144070669" observedRunningTime="2025-10-01 16:10:53.827465268 +0000 UTC m=+706.729017855" watchObservedRunningTime="2025-10-01 16:10:53.828911741 +0000 UTC m=+706.730464328" Oct 01 16:10:55 crc kubenswrapper[4726]: I1001 16:10:55.782157 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" event={"ID":"f179fe18-0a34-43df-bb39-dab69a9ee29b","Type":"ContainerStarted","Data":"41349d4807fad2f1eb449c5da33539c938175fdc87d92111b1019282e17302a8"} Oct 01 16:10:55 crc kubenswrapper[4726]: I1001 16:10:55.803550 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" podStartSLOduration=1.468977782 podStartE2EDuration="6.803525952s" podCreationTimestamp="2025-10-01 16:10:49 +0000 UTC" firstStartedPulling="2025-10-01 16:10:50.272922015 +0000 UTC m=+703.174474602" lastFinishedPulling="2025-10-01 16:10:55.607470195 +0000 UTC m=+708.509022772" observedRunningTime="2025-10-01 16:10:55.799832313 +0000 UTC m=+708.701384910" watchObservedRunningTime="2025-10-01 16:10:55.803525952 +0000 UTC m=+708.705078569" Oct 01 16:11:00 crc kubenswrapper[4726]: I1001 16:11:00.077756 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:11:00 crc kubenswrapper[4726]: I1001 16:11:00.340941 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:11:00 crc kubenswrapper[4726]: I1001 16:11:00.341029 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:11:00 crc kubenswrapper[4726]: I1001 16:11:00.348877 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:11:00 crc kubenswrapper[4726]: I1001 16:11:00.825388 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5f4d7b984b-cpbsb" Oct 01 16:11:00 crc kubenswrapper[4726]: I1001 16:11:00.909265 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-w4c9z"] Oct 01 16:11:10 crc kubenswrapper[4726]: I1001 16:11:10.621116 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" Oct 01 16:11:23 crc kubenswrapper[4726]: I1001 16:11:23.413335 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:11:23 crc kubenswrapper[4726]: I1001 16:11:23.414089 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:11:24 crc kubenswrapper[4726]: I1001 16:11:24.975665 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b"] Oct 01 16:11:24 crc kubenswrapper[4726]: I1001 16:11:24.977482 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" Oct 01 16:11:24 crc kubenswrapper[4726]: I1001 16:11:24.980325 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 01 16:11:24 crc kubenswrapper[4726]: I1001 16:11:24.991419 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b"] Oct 01 16:11:25 crc kubenswrapper[4726]: I1001 16:11:25.067321 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b751f5f8-ea87-4fcb-8c07-93863ae71d03-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b\" (UID: \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" Oct 01 16:11:25 crc kubenswrapper[4726]: I1001 16:11:25.067668 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5w24\" (UniqueName: \"kubernetes.io/projected/b751f5f8-ea87-4fcb-8c07-93863ae71d03-kube-api-access-h5w24\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b\" (UID: \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" Oct 01 16:11:25 crc kubenswrapper[4726]: I1001 16:11:25.067883 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b751f5f8-ea87-4fcb-8c07-93863ae71d03-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b\" (UID: \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" Oct 01 16:11:25 crc kubenswrapper[4726]: I1001 16:11:25.169003 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5w24\" (UniqueName: \"kubernetes.io/projected/b751f5f8-ea87-4fcb-8c07-93863ae71d03-kube-api-access-h5w24\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b\" (UID: \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" Oct 01 16:11:25 crc kubenswrapper[4726]: I1001 16:11:25.169138 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b751f5f8-ea87-4fcb-8c07-93863ae71d03-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b\" (UID: \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" Oct 01 16:11:25 crc kubenswrapper[4726]: I1001 16:11:25.169211 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b751f5f8-ea87-4fcb-8c07-93863ae71d03-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b\" (UID: \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" Oct 01 16:11:25 crc kubenswrapper[4726]: I1001 16:11:25.170489 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b751f5f8-ea87-4fcb-8c07-93863ae71d03-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b\" (UID: \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" Oct 01 16:11:25 crc kubenswrapper[4726]: I1001 16:11:25.170490 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b751f5f8-ea87-4fcb-8c07-93863ae71d03-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b\" (UID: \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" Oct 01 16:11:25 crc kubenswrapper[4726]: I1001 16:11:25.204745 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5w24\" (UniqueName: \"kubernetes.io/projected/b751f5f8-ea87-4fcb-8c07-93863ae71d03-kube-api-access-h5w24\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b\" (UID: \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" Oct 01 16:11:25 crc kubenswrapper[4726]: I1001 16:11:25.305817 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" Oct 01 16:11:25 crc kubenswrapper[4726]: I1001 16:11:25.774496 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b"] Oct 01 16:11:25 crc kubenswrapper[4726]: I1001 16:11:25.990384 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" event={"ID":"b751f5f8-ea87-4fcb-8c07-93863ae71d03","Type":"ContainerStarted","Data":"1369ec55c12455cf89b174acf28f9a7910660154de5342ae1a71afb2f808c650"} Oct 01 16:11:25 crc kubenswrapper[4726]: I1001 16:11:25.993979 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-w4c9z" podUID="79d896b5-b34d-4206-92d8-99caeafdffb2" containerName="console" containerID="cri-o://462d230384168689aa43fc8e1ca8f9835c9bd43481602a2559c9d7c9f58e35ab" gracePeriod=15 Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.360060 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-w4c9z_79d896b5-b34d-4206-92d8-99caeafdffb2/console/0.log" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.360355 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.490266 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-trusted-ca-bundle\") pod \"79d896b5-b34d-4206-92d8-99caeafdffb2\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.490331 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/79d896b5-b34d-4206-92d8-99caeafdffb2-console-serving-cert\") pod \"79d896b5-b34d-4206-92d8-99caeafdffb2\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.490351 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/79d896b5-b34d-4206-92d8-99caeafdffb2-console-oauth-config\") pod \"79d896b5-b34d-4206-92d8-99caeafdffb2\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.490380 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-service-ca\") pod \"79d896b5-b34d-4206-92d8-99caeafdffb2\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.490462 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-oauth-serving-cert\") pod \"79d896b5-b34d-4206-92d8-99caeafdffb2\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.491198 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-service-ca" (OuterVolumeSpecName: "service-ca") pod "79d896b5-b34d-4206-92d8-99caeafdffb2" (UID: "79d896b5-b34d-4206-92d8-99caeafdffb2"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.491234 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x82vr\" (UniqueName: \"kubernetes.io/projected/79d896b5-b34d-4206-92d8-99caeafdffb2-kube-api-access-x82vr\") pod \"79d896b5-b34d-4206-92d8-99caeafdffb2\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.491270 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-console-config\") pod \"79d896b5-b34d-4206-92d8-99caeafdffb2\" (UID: \"79d896b5-b34d-4206-92d8-99caeafdffb2\") " Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.491277 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "79d896b5-b34d-4206-92d8-99caeafdffb2" (UID: "79d896b5-b34d-4206-92d8-99caeafdffb2"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.491428 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "79d896b5-b34d-4206-92d8-99caeafdffb2" (UID: "79d896b5-b34d-4206-92d8-99caeafdffb2"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.491714 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-console-config" (OuterVolumeSpecName: "console-config") pod "79d896b5-b34d-4206-92d8-99caeafdffb2" (UID: "79d896b5-b34d-4206-92d8-99caeafdffb2"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.491880 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.491893 4726 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.491903 4726 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-console-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.491912 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/79d896b5-b34d-4206-92d8-99caeafdffb2-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.498119 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79d896b5-b34d-4206-92d8-99caeafdffb2-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "79d896b5-b34d-4206-92d8-99caeafdffb2" (UID: "79d896b5-b34d-4206-92d8-99caeafdffb2"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.498449 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79d896b5-b34d-4206-92d8-99caeafdffb2-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "79d896b5-b34d-4206-92d8-99caeafdffb2" (UID: "79d896b5-b34d-4206-92d8-99caeafdffb2"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.498491 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79d896b5-b34d-4206-92d8-99caeafdffb2-kube-api-access-x82vr" (OuterVolumeSpecName: "kube-api-access-x82vr") pod "79d896b5-b34d-4206-92d8-99caeafdffb2" (UID: "79d896b5-b34d-4206-92d8-99caeafdffb2"). InnerVolumeSpecName "kube-api-access-x82vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.592809 4726 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/79d896b5-b34d-4206-92d8-99caeafdffb2-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.592841 4726 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/79d896b5-b34d-4206-92d8-99caeafdffb2-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:26 crc kubenswrapper[4726]: I1001 16:11:26.592854 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x82vr\" (UniqueName: \"kubernetes.io/projected/79d896b5-b34d-4206-92d8-99caeafdffb2-kube-api-access-x82vr\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:27 crc kubenswrapper[4726]: I1001 16:11:27.015480 4726 generic.go:334] "Generic (PLEG): container finished" podID="b751f5f8-ea87-4fcb-8c07-93863ae71d03" containerID="83e7f4337aebfc23a8b97e98f34f115ca4d34817d97c1f5b045118dcd4cc963b" exitCode=0 Oct 01 16:11:27 crc kubenswrapper[4726]: I1001 16:11:27.015617 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" event={"ID":"b751f5f8-ea87-4fcb-8c07-93863ae71d03","Type":"ContainerDied","Data":"83e7f4337aebfc23a8b97e98f34f115ca4d34817d97c1f5b045118dcd4cc963b"} Oct 01 16:11:27 crc kubenswrapper[4726]: I1001 16:11:27.018904 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-w4c9z_79d896b5-b34d-4206-92d8-99caeafdffb2/console/0.log" Oct 01 16:11:27 crc kubenswrapper[4726]: I1001 16:11:27.018958 4726 generic.go:334] "Generic (PLEG): container finished" podID="79d896b5-b34d-4206-92d8-99caeafdffb2" containerID="462d230384168689aa43fc8e1ca8f9835c9bd43481602a2559c9d7c9f58e35ab" exitCode=2 Oct 01 16:11:27 crc kubenswrapper[4726]: I1001 16:11:27.018989 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-w4c9z" event={"ID":"79d896b5-b34d-4206-92d8-99caeafdffb2","Type":"ContainerDied","Data":"462d230384168689aa43fc8e1ca8f9835c9bd43481602a2559c9d7c9f58e35ab"} Oct 01 16:11:27 crc kubenswrapper[4726]: I1001 16:11:27.019016 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-w4c9z" event={"ID":"79d896b5-b34d-4206-92d8-99caeafdffb2","Type":"ContainerDied","Data":"86c2278e12d739cff7dc0245e601e8400ceb9f8d6cd49937c6ef877acd41a21b"} Oct 01 16:11:27 crc kubenswrapper[4726]: I1001 16:11:27.019035 4726 scope.go:117] "RemoveContainer" containerID="462d230384168689aa43fc8e1ca8f9835c9bd43481602a2559c9d7c9f58e35ab" Oct 01 16:11:27 crc kubenswrapper[4726]: I1001 16:11:27.020466 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-w4c9z" Oct 01 16:11:27 crc kubenswrapper[4726]: I1001 16:11:27.064363 4726 scope.go:117] "RemoveContainer" containerID="462d230384168689aa43fc8e1ca8f9835c9bd43481602a2559c9d7c9f58e35ab" Oct 01 16:11:27 crc kubenswrapper[4726]: E1001 16:11:27.065187 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"462d230384168689aa43fc8e1ca8f9835c9bd43481602a2559c9d7c9f58e35ab\": container with ID starting with 462d230384168689aa43fc8e1ca8f9835c9bd43481602a2559c9d7c9f58e35ab not found: ID does not exist" containerID="462d230384168689aa43fc8e1ca8f9835c9bd43481602a2559c9d7c9f58e35ab" Oct 01 16:11:27 crc kubenswrapper[4726]: I1001 16:11:27.065773 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"462d230384168689aa43fc8e1ca8f9835c9bd43481602a2559c9d7c9f58e35ab"} err="failed to get container status \"462d230384168689aa43fc8e1ca8f9835c9bd43481602a2559c9d7c9f58e35ab\": rpc error: code = NotFound desc = could not find container \"462d230384168689aa43fc8e1ca8f9835c9bd43481602a2559c9d7c9f58e35ab\": container with ID starting with 462d230384168689aa43fc8e1ca8f9835c9bd43481602a2559c9d7c9f58e35ab not found: ID does not exist" Oct 01 16:11:27 crc kubenswrapper[4726]: I1001 16:11:27.074174 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-w4c9z"] Oct 01 16:11:27 crc kubenswrapper[4726]: I1001 16:11:27.079636 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-w4c9z"] Oct 01 16:11:27 crc kubenswrapper[4726]: I1001 16:11:27.817232 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79d896b5-b34d-4206-92d8-99caeafdffb2" path="/var/lib/kubelet/pods/79d896b5-b34d-4206-92d8-99caeafdffb2/volumes" Oct 01 16:11:29 crc kubenswrapper[4726]: I1001 16:11:29.040192 4726 generic.go:334] "Generic (PLEG): container finished" podID="b751f5f8-ea87-4fcb-8c07-93863ae71d03" containerID="a5d82b1bdd886813154970da17f6fda950d6dd7b66037e813ee9e138ee769893" exitCode=0 Oct 01 16:11:29 crc kubenswrapper[4726]: I1001 16:11:29.040280 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" event={"ID":"b751f5f8-ea87-4fcb-8c07-93863ae71d03","Type":"ContainerDied","Data":"a5d82b1bdd886813154970da17f6fda950d6dd7b66037e813ee9e138ee769893"} Oct 01 16:11:30 crc kubenswrapper[4726]: I1001 16:11:30.050493 4726 generic.go:334] "Generic (PLEG): container finished" podID="b751f5f8-ea87-4fcb-8c07-93863ae71d03" containerID="b53cd33d02e122b1528469a75fd3318c29361c739978b2402520f51c10e823f1" exitCode=0 Oct 01 16:11:30 crc kubenswrapper[4726]: I1001 16:11:30.050551 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" event={"ID":"b751f5f8-ea87-4fcb-8c07-93863ae71d03","Type":"ContainerDied","Data":"b53cd33d02e122b1528469a75fd3318c29361c739978b2402520f51c10e823f1"} Oct 01 16:11:31 crc kubenswrapper[4726]: I1001 16:11:31.345143 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" Oct 01 16:11:31 crc kubenswrapper[4726]: I1001 16:11:31.468957 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b751f5f8-ea87-4fcb-8c07-93863ae71d03-bundle\") pod \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\" (UID: \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\") " Oct 01 16:11:31 crc kubenswrapper[4726]: I1001 16:11:31.469004 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5w24\" (UniqueName: \"kubernetes.io/projected/b751f5f8-ea87-4fcb-8c07-93863ae71d03-kube-api-access-h5w24\") pod \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\" (UID: \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\") " Oct 01 16:11:31 crc kubenswrapper[4726]: I1001 16:11:31.469068 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b751f5f8-ea87-4fcb-8c07-93863ae71d03-util\") pod \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\" (UID: \"b751f5f8-ea87-4fcb-8c07-93863ae71d03\") " Oct 01 16:11:31 crc kubenswrapper[4726]: I1001 16:11:31.470143 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b751f5f8-ea87-4fcb-8c07-93863ae71d03-bundle" (OuterVolumeSpecName: "bundle") pod "b751f5f8-ea87-4fcb-8c07-93863ae71d03" (UID: "b751f5f8-ea87-4fcb-8c07-93863ae71d03"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:11:31 crc kubenswrapper[4726]: I1001 16:11:31.484340 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b751f5f8-ea87-4fcb-8c07-93863ae71d03-kube-api-access-h5w24" (OuterVolumeSpecName: "kube-api-access-h5w24") pod "b751f5f8-ea87-4fcb-8c07-93863ae71d03" (UID: "b751f5f8-ea87-4fcb-8c07-93863ae71d03"). InnerVolumeSpecName "kube-api-access-h5w24". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:11:31 crc kubenswrapper[4726]: I1001 16:11:31.570676 4726 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b751f5f8-ea87-4fcb-8c07-93863ae71d03-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:31 crc kubenswrapper[4726]: I1001 16:11:31.570712 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5w24\" (UniqueName: \"kubernetes.io/projected/b751f5f8-ea87-4fcb-8c07-93863ae71d03-kube-api-access-h5w24\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:31 crc kubenswrapper[4726]: I1001 16:11:31.589597 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b751f5f8-ea87-4fcb-8c07-93863ae71d03-util" (OuterVolumeSpecName: "util") pod "b751f5f8-ea87-4fcb-8c07-93863ae71d03" (UID: "b751f5f8-ea87-4fcb-8c07-93863ae71d03"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:11:31 crc kubenswrapper[4726]: I1001 16:11:31.672265 4726 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b751f5f8-ea87-4fcb-8c07-93863ae71d03-util\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:32 crc kubenswrapper[4726]: I1001 16:11:32.068728 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" event={"ID":"b751f5f8-ea87-4fcb-8c07-93863ae71d03","Type":"ContainerDied","Data":"1369ec55c12455cf89b174acf28f9a7910660154de5342ae1a71afb2f808c650"} Oct 01 16:11:32 crc kubenswrapper[4726]: I1001 16:11:32.068797 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1369ec55c12455cf89b174acf28f9a7910660154de5342ae1a71afb2f808c650" Oct 01 16:11:32 crc kubenswrapper[4726]: I1001 16:11:32.068829 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.538644 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh"] Oct 01 16:11:41 crc kubenswrapper[4726]: E1001 16:11:41.539406 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79d896b5-b34d-4206-92d8-99caeafdffb2" containerName="console" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.539418 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="79d896b5-b34d-4206-92d8-99caeafdffb2" containerName="console" Oct 01 16:11:41 crc kubenswrapper[4726]: E1001 16:11:41.539435 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b751f5f8-ea87-4fcb-8c07-93863ae71d03" containerName="util" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.539441 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b751f5f8-ea87-4fcb-8c07-93863ae71d03" containerName="util" Oct 01 16:11:41 crc kubenswrapper[4726]: E1001 16:11:41.539452 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b751f5f8-ea87-4fcb-8c07-93863ae71d03" containerName="extract" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.539458 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b751f5f8-ea87-4fcb-8c07-93863ae71d03" containerName="extract" Oct 01 16:11:41 crc kubenswrapper[4726]: E1001 16:11:41.539468 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b751f5f8-ea87-4fcb-8c07-93863ae71d03" containerName="pull" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.539473 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b751f5f8-ea87-4fcb-8c07-93863ae71d03" containerName="pull" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.539578 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="79d896b5-b34d-4206-92d8-99caeafdffb2" containerName="console" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.539592 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b751f5f8-ea87-4fcb-8c07-93863ae71d03" containerName="extract" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.539979 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.543345 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.543690 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.543729 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-fs5pn" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.543987 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.554434 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.564664 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh"] Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.698607 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94e8d60f-8f3a-448a-9096-163826d5ad92-webhook-cert\") pod \"metallb-operator-controller-manager-86d5dd6867-8btsh\" (UID: \"94e8d60f-8f3a-448a-9096-163826d5ad92\") " pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.698664 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khxpb\" (UniqueName: \"kubernetes.io/projected/94e8d60f-8f3a-448a-9096-163826d5ad92-kube-api-access-khxpb\") pod \"metallb-operator-controller-manager-86d5dd6867-8btsh\" (UID: \"94e8d60f-8f3a-448a-9096-163826d5ad92\") " pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.698685 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94e8d60f-8f3a-448a-9096-163826d5ad92-apiservice-cert\") pod \"metallb-operator-controller-manager-86d5dd6867-8btsh\" (UID: \"94e8d60f-8f3a-448a-9096-163826d5ad92\") " pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.800222 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94e8d60f-8f3a-448a-9096-163826d5ad92-webhook-cert\") pod \"metallb-operator-controller-manager-86d5dd6867-8btsh\" (UID: \"94e8d60f-8f3a-448a-9096-163826d5ad92\") " pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.800606 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khxpb\" (UniqueName: \"kubernetes.io/projected/94e8d60f-8f3a-448a-9096-163826d5ad92-kube-api-access-khxpb\") pod \"metallb-operator-controller-manager-86d5dd6867-8btsh\" (UID: \"94e8d60f-8f3a-448a-9096-163826d5ad92\") " pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.800711 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94e8d60f-8f3a-448a-9096-163826d5ad92-apiservice-cert\") pod \"metallb-operator-controller-manager-86d5dd6867-8btsh\" (UID: \"94e8d60f-8f3a-448a-9096-163826d5ad92\") " pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.806913 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94e8d60f-8f3a-448a-9096-163826d5ad92-webhook-cert\") pod \"metallb-operator-controller-manager-86d5dd6867-8btsh\" (UID: \"94e8d60f-8f3a-448a-9096-163826d5ad92\") " pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.809620 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94e8d60f-8f3a-448a-9096-163826d5ad92-apiservice-cert\") pod \"metallb-operator-controller-manager-86d5dd6867-8btsh\" (UID: \"94e8d60f-8f3a-448a-9096-163826d5ad92\") " pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.817204 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khxpb\" (UniqueName: \"kubernetes.io/projected/94e8d60f-8f3a-448a-9096-163826d5ad92-kube-api-access-khxpb\") pod \"metallb-operator-controller-manager-86d5dd6867-8btsh\" (UID: \"94e8d60f-8f3a-448a-9096-163826d5ad92\") " pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.855418 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.880659 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48"] Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.881941 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.884089 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.884607 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.884937 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-c8r2p" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.901871 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j66rg\" (UniqueName: \"kubernetes.io/projected/daa5265d-09f5-4a83-96a0-415448d42bcd-kube-api-access-j66rg\") pod \"metallb-operator-webhook-server-d69b94c67-gnx48\" (UID: \"daa5265d-09f5-4a83-96a0-415448d42bcd\") " pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.901950 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/daa5265d-09f5-4a83-96a0-415448d42bcd-apiservice-cert\") pod \"metallb-operator-webhook-server-d69b94c67-gnx48\" (UID: \"daa5265d-09f5-4a83-96a0-415448d42bcd\") " pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.901977 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/daa5265d-09f5-4a83-96a0-415448d42bcd-webhook-cert\") pod \"metallb-operator-webhook-server-d69b94c67-gnx48\" (UID: \"daa5265d-09f5-4a83-96a0-415448d42bcd\") " pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:11:41 crc kubenswrapper[4726]: I1001 16:11:41.941416 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48"] Oct 01 16:11:42 crc kubenswrapper[4726]: I1001 16:11:42.004596 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/daa5265d-09f5-4a83-96a0-415448d42bcd-apiservice-cert\") pod \"metallb-operator-webhook-server-d69b94c67-gnx48\" (UID: \"daa5265d-09f5-4a83-96a0-415448d42bcd\") " pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:11:42 crc kubenswrapper[4726]: I1001 16:11:42.004886 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/daa5265d-09f5-4a83-96a0-415448d42bcd-webhook-cert\") pod \"metallb-operator-webhook-server-d69b94c67-gnx48\" (UID: \"daa5265d-09f5-4a83-96a0-415448d42bcd\") " pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:11:42 crc kubenswrapper[4726]: I1001 16:11:42.004964 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j66rg\" (UniqueName: \"kubernetes.io/projected/daa5265d-09f5-4a83-96a0-415448d42bcd-kube-api-access-j66rg\") pod \"metallb-operator-webhook-server-d69b94c67-gnx48\" (UID: \"daa5265d-09f5-4a83-96a0-415448d42bcd\") " pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:11:42 crc kubenswrapper[4726]: I1001 16:11:42.011342 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/daa5265d-09f5-4a83-96a0-415448d42bcd-apiservice-cert\") pod \"metallb-operator-webhook-server-d69b94c67-gnx48\" (UID: \"daa5265d-09f5-4a83-96a0-415448d42bcd\") " pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:11:42 crc kubenswrapper[4726]: I1001 16:11:42.011981 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/daa5265d-09f5-4a83-96a0-415448d42bcd-webhook-cert\") pod \"metallb-operator-webhook-server-d69b94c67-gnx48\" (UID: \"daa5265d-09f5-4a83-96a0-415448d42bcd\") " pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:11:42 crc kubenswrapper[4726]: I1001 16:11:42.032830 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j66rg\" (UniqueName: \"kubernetes.io/projected/daa5265d-09f5-4a83-96a0-415448d42bcd-kube-api-access-j66rg\") pod \"metallb-operator-webhook-server-d69b94c67-gnx48\" (UID: \"daa5265d-09f5-4a83-96a0-415448d42bcd\") " pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:11:42 crc kubenswrapper[4726]: I1001 16:11:42.134509 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh"] Oct 01 16:11:42 crc kubenswrapper[4726]: I1001 16:11:42.235960 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:11:42 crc kubenswrapper[4726]: I1001 16:11:42.499997 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48"] Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.135872 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" event={"ID":"daa5265d-09f5-4a83-96a0-415448d42bcd","Type":"ContainerStarted","Data":"e8a5c4ec52f29295eb52c16adb0e2fc48767851c884aa45893448252ce2564af"} Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.138601 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" event={"ID":"94e8d60f-8f3a-448a-9096-163826d5ad92","Type":"ContainerStarted","Data":"c10990bc3975e695e6ded271396d6eb2fd0ed59d64ec23c6caf3c82be4bdb321"} Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.210775 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zhfrh"] Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.210985 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" podUID="73789e0a-a83d-4470-9f4c-137b3c317b85" containerName="controller-manager" containerID="cri-o://1130beea1319e8fa0fc79c765ffa78d587a366351938b179f4cc04bd68f8238e" gracePeriod=30 Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.232472 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8"] Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.232728 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" podUID="1a8d300c-d03d-4927-b9ba-c17f692a9db4" containerName="route-controller-manager" containerID="cri-o://dfb16ccd3f91ddf9f6fd4194242c4a779dc999ad3abc3556c4666a4c4ed2139d" gracePeriod=30 Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.569692 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.613358 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.725716 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-config\") pod \"73789e0a-a83d-4470-9f4c-137b3c317b85\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.725765 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-proxy-ca-bundles\") pod \"73789e0a-a83d-4470-9f4c-137b3c317b85\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.725781 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-client-ca\") pod \"73789e0a-a83d-4470-9f4c-137b3c317b85\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.725810 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nt7mt\" (UniqueName: \"kubernetes.io/projected/1a8d300c-d03d-4927-b9ba-c17f692a9db4-kube-api-access-nt7mt\") pod \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.725846 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a8d300c-d03d-4927-b9ba-c17f692a9db4-serving-cert\") pod \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.725899 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73789e0a-a83d-4470-9f4c-137b3c317b85-serving-cert\") pod \"73789e0a-a83d-4470-9f4c-137b3c317b85\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.725924 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a8d300c-d03d-4927-b9ba-c17f692a9db4-client-ca\") pod \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.725962 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a8d300c-d03d-4927-b9ba-c17f692a9db4-config\") pod \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\" (UID: \"1a8d300c-d03d-4927-b9ba-c17f692a9db4\") " Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.725978 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27mwc\" (UniqueName: \"kubernetes.io/projected/73789e0a-a83d-4470-9f4c-137b3c317b85-kube-api-access-27mwc\") pod \"73789e0a-a83d-4470-9f4c-137b3c317b85\" (UID: \"73789e0a-a83d-4470-9f4c-137b3c317b85\") " Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.728117 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a8d300c-d03d-4927-b9ba-c17f692a9db4-client-ca" (OuterVolumeSpecName: "client-ca") pod "1a8d300c-d03d-4927-b9ba-c17f692a9db4" (UID: "1a8d300c-d03d-4927-b9ba-c17f692a9db4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.728237 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a8d300c-d03d-4927-b9ba-c17f692a9db4-config" (OuterVolumeSpecName: "config") pod "1a8d300c-d03d-4927-b9ba-c17f692a9db4" (UID: "1a8d300c-d03d-4927-b9ba-c17f692a9db4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.728558 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "73789e0a-a83d-4470-9f4c-137b3c317b85" (UID: "73789e0a-a83d-4470-9f4c-137b3c317b85"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.728667 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-client-ca" (OuterVolumeSpecName: "client-ca") pod "73789e0a-a83d-4470-9f4c-137b3c317b85" (UID: "73789e0a-a83d-4470-9f4c-137b3c317b85"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.728994 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-config" (OuterVolumeSpecName: "config") pod "73789e0a-a83d-4470-9f4c-137b3c317b85" (UID: "73789e0a-a83d-4470-9f4c-137b3c317b85"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.736209 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73789e0a-a83d-4470-9f4c-137b3c317b85-kube-api-access-27mwc" (OuterVolumeSpecName: "kube-api-access-27mwc") pod "73789e0a-a83d-4470-9f4c-137b3c317b85" (UID: "73789e0a-a83d-4470-9f4c-137b3c317b85"). InnerVolumeSpecName "kube-api-access-27mwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.736433 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73789e0a-a83d-4470-9f4c-137b3c317b85-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "73789e0a-a83d-4470-9f4c-137b3c317b85" (UID: "73789e0a-a83d-4470-9f4c-137b3c317b85"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.736557 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a8d300c-d03d-4927-b9ba-c17f692a9db4-kube-api-access-nt7mt" (OuterVolumeSpecName: "kube-api-access-nt7mt") pod "1a8d300c-d03d-4927-b9ba-c17f692a9db4" (UID: "1a8d300c-d03d-4927-b9ba-c17f692a9db4"). InnerVolumeSpecName "kube-api-access-nt7mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.736597 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a8d300c-d03d-4927-b9ba-c17f692a9db4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1a8d300c-d03d-4927-b9ba-c17f692a9db4" (UID: "1a8d300c-d03d-4927-b9ba-c17f692a9db4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.759125 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-775b8bd7f8-45s27"] Oct 01 16:11:43 crc kubenswrapper[4726]: E1001 16:11:43.759379 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73789e0a-a83d-4470-9f4c-137b3c317b85" containerName="controller-manager" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.759398 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="73789e0a-a83d-4470-9f4c-137b3c317b85" containerName="controller-manager" Oct 01 16:11:43 crc kubenswrapper[4726]: E1001 16:11:43.759412 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a8d300c-d03d-4927-b9ba-c17f692a9db4" containerName="route-controller-manager" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.759421 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a8d300c-d03d-4927-b9ba-c17f692a9db4" containerName="route-controller-manager" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.759533 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="73789e0a-a83d-4470-9f4c-137b3c317b85" containerName="controller-manager" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.759550 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a8d300c-d03d-4927-b9ba-c17f692a9db4" containerName="route-controller-manager" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.759952 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.769243 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-775b8bd7f8-45s27"] Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.830277 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73789e0a-a83d-4470-9f4c-137b3c317b85-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.830307 4726 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a8d300c-d03d-4927-b9ba-c17f692a9db4-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.830325 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a8d300c-d03d-4927-b9ba-c17f692a9db4-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.830335 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27mwc\" (UniqueName: \"kubernetes.io/projected/73789e0a-a83d-4470-9f4c-137b3c317b85-kube-api-access-27mwc\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.830346 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.830355 4726 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.830363 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/73789e0a-a83d-4470-9f4c-137b3c317b85-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.830372 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nt7mt\" (UniqueName: \"kubernetes.io/projected/1a8d300c-d03d-4927-b9ba-c17f692a9db4-kube-api-access-nt7mt\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.830381 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a8d300c-d03d-4927-b9ba-c17f692a9db4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.849409 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n"] Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.850077 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.862765 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n"] Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.931547 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9bcab5a3-6477-47d4-a954-de6f55771a61-proxy-ca-bundles\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.931753 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bcab5a3-6477-47d4-a954-de6f55771a61-client-ca\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.931804 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkb64\" (UniqueName: \"kubernetes.io/projected/9bcab5a3-6477-47d4-a954-de6f55771a61-kube-api-access-qkb64\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.932380 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bcab5a3-6477-47d4-a954-de6f55771a61-serving-cert\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:43 crc kubenswrapper[4726]: I1001 16:11:43.932786 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bcab5a3-6477-47d4-a954-de6f55771a61-config\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.034522 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlzwz\" (UniqueName: \"kubernetes.io/projected/0438d3a0-491c-40b3-b596-1940a287b882-kube-api-access-hlzwz\") pod \"route-controller-manager-74dbd8bd9b-jz74n\" (UID: \"0438d3a0-491c-40b3-b596-1940a287b882\") " pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.034589 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bcab5a3-6477-47d4-a954-de6f55771a61-client-ca\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.034666 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkb64\" (UniqueName: \"kubernetes.io/projected/9bcab5a3-6477-47d4-a954-de6f55771a61-kube-api-access-qkb64\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.034713 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0438d3a0-491c-40b3-b596-1940a287b882-client-ca\") pod \"route-controller-manager-74dbd8bd9b-jz74n\" (UID: \"0438d3a0-491c-40b3-b596-1940a287b882\") " pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.034739 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bcab5a3-6477-47d4-a954-de6f55771a61-serving-cert\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.034771 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bcab5a3-6477-47d4-a954-de6f55771a61-config\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.034802 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0438d3a0-491c-40b3-b596-1940a287b882-config\") pod \"route-controller-manager-74dbd8bd9b-jz74n\" (UID: \"0438d3a0-491c-40b3-b596-1940a287b882\") " pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.034823 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9bcab5a3-6477-47d4-a954-de6f55771a61-proxy-ca-bundles\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.034848 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0438d3a0-491c-40b3-b596-1940a287b882-serving-cert\") pod \"route-controller-manager-74dbd8bd9b-jz74n\" (UID: \"0438d3a0-491c-40b3-b596-1940a287b882\") " pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.035492 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bcab5a3-6477-47d4-a954-de6f55771a61-client-ca\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.037347 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bcab5a3-6477-47d4-a954-de6f55771a61-config\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.037446 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9bcab5a3-6477-47d4-a954-de6f55771a61-proxy-ca-bundles\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.050181 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bcab5a3-6477-47d4-a954-de6f55771a61-serving-cert\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.054844 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkb64\" (UniqueName: \"kubernetes.io/projected/9bcab5a3-6477-47d4-a954-de6f55771a61-kube-api-access-qkb64\") pod \"controller-manager-775b8bd7f8-45s27\" (UID: \"9bcab5a3-6477-47d4-a954-de6f55771a61\") " pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.078885 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.135688 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0438d3a0-491c-40b3-b596-1940a287b882-client-ca\") pod \"route-controller-manager-74dbd8bd9b-jz74n\" (UID: \"0438d3a0-491c-40b3-b596-1940a287b882\") " pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.135884 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0438d3a0-491c-40b3-b596-1940a287b882-config\") pod \"route-controller-manager-74dbd8bd9b-jz74n\" (UID: \"0438d3a0-491c-40b3-b596-1940a287b882\") " pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.135915 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0438d3a0-491c-40b3-b596-1940a287b882-serving-cert\") pod \"route-controller-manager-74dbd8bd9b-jz74n\" (UID: \"0438d3a0-491c-40b3-b596-1940a287b882\") " pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.135956 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlzwz\" (UniqueName: \"kubernetes.io/projected/0438d3a0-491c-40b3-b596-1940a287b882-kube-api-access-hlzwz\") pod \"route-controller-manager-74dbd8bd9b-jz74n\" (UID: \"0438d3a0-491c-40b3-b596-1940a287b882\") " pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.137689 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0438d3a0-491c-40b3-b596-1940a287b882-client-ca\") pod \"route-controller-manager-74dbd8bd9b-jz74n\" (UID: \"0438d3a0-491c-40b3-b596-1940a287b882\") " pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.139639 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0438d3a0-491c-40b3-b596-1940a287b882-config\") pod \"route-controller-manager-74dbd8bd9b-jz74n\" (UID: \"0438d3a0-491c-40b3-b596-1940a287b882\") " pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.142314 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0438d3a0-491c-40b3-b596-1940a287b882-serving-cert\") pod \"route-controller-manager-74dbd8bd9b-jz74n\" (UID: \"0438d3a0-491c-40b3-b596-1940a287b882\") " pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.146780 4726 generic.go:334] "Generic (PLEG): container finished" podID="1a8d300c-d03d-4927-b9ba-c17f692a9db4" containerID="dfb16ccd3f91ddf9f6fd4194242c4a779dc999ad3abc3556c4666a4c4ed2139d" exitCode=0 Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.146825 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" event={"ID":"1a8d300c-d03d-4927-b9ba-c17f692a9db4","Type":"ContainerDied","Data":"dfb16ccd3f91ddf9f6fd4194242c4a779dc999ad3abc3556c4666a4c4ed2139d"} Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.146882 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" event={"ID":"1a8d300c-d03d-4927-b9ba-c17f692a9db4","Type":"ContainerDied","Data":"f6fcd61b312ae9185504efa02bc5a3fe03cbc70d476de06e6702a6037821bcb5"} Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.146900 4726 scope.go:117] "RemoveContainer" containerID="dfb16ccd3f91ddf9f6fd4194242c4a779dc999ad3abc3556c4666a4c4ed2139d" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.147075 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.155793 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.156045 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" event={"ID":"73789e0a-a83d-4470-9f4c-137b3c317b85","Type":"ContainerDied","Data":"1130beea1319e8fa0fc79c765ffa78d587a366351938b179f4cc04bd68f8238e"} Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.155679 4726 generic.go:334] "Generic (PLEG): container finished" podID="73789e0a-a83d-4470-9f4c-137b3c317b85" containerID="1130beea1319e8fa0fc79c765ffa78d587a366351938b179f4cc04bd68f8238e" exitCode=0 Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.156242 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zhfrh" event={"ID":"73789e0a-a83d-4470-9f4c-137b3c317b85","Type":"ContainerDied","Data":"28894c50be0b5ef619576dd433216ee43b3d48b8cf5f6b9065b21a32fb43d9fb"} Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.156900 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlzwz\" (UniqueName: \"kubernetes.io/projected/0438d3a0-491c-40b3-b596-1940a287b882-kube-api-access-hlzwz\") pod \"route-controller-manager-74dbd8bd9b-jz74n\" (UID: \"0438d3a0-491c-40b3-b596-1940a287b882\") " pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.165462 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8"] Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.165526 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.170381 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2gcb8"] Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.176978 4726 scope.go:117] "RemoveContainer" containerID="dfb16ccd3f91ddf9f6fd4194242c4a779dc999ad3abc3556c4666a4c4ed2139d" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.177093 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zhfrh"] Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.180582 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zhfrh"] Oct 01 16:11:44 crc kubenswrapper[4726]: E1001 16:11:44.184993 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfb16ccd3f91ddf9f6fd4194242c4a779dc999ad3abc3556c4666a4c4ed2139d\": container with ID starting with dfb16ccd3f91ddf9f6fd4194242c4a779dc999ad3abc3556c4666a4c4ed2139d not found: ID does not exist" containerID="dfb16ccd3f91ddf9f6fd4194242c4a779dc999ad3abc3556c4666a4c4ed2139d" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.185024 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfb16ccd3f91ddf9f6fd4194242c4a779dc999ad3abc3556c4666a4c4ed2139d"} err="failed to get container status \"dfb16ccd3f91ddf9f6fd4194242c4a779dc999ad3abc3556c4666a4c4ed2139d\": rpc error: code = NotFound desc = could not find container \"dfb16ccd3f91ddf9f6fd4194242c4a779dc999ad3abc3556c4666a4c4ed2139d\": container with ID starting with dfb16ccd3f91ddf9f6fd4194242c4a779dc999ad3abc3556c4666a4c4ed2139d not found: ID does not exist" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.185060 4726 scope.go:117] "RemoveContainer" containerID="1130beea1319e8fa0fc79c765ffa78d587a366351938b179f4cc04bd68f8238e" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.231222 4726 scope.go:117] "RemoveContainer" containerID="1130beea1319e8fa0fc79c765ffa78d587a366351938b179f4cc04bd68f8238e" Oct 01 16:11:44 crc kubenswrapper[4726]: E1001 16:11:44.239190 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1130beea1319e8fa0fc79c765ffa78d587a366351938b179f4cc04bd68f8238e\": container with ID starting with 1130beea1319e8fa0fc79c765ffa78d587a366351938b179f4cc04bd68f8238e not found: ID does not exist" containerID="1130beea1319e8fa0fc79c765ffa78d587a366351938b179f4cc04bd68f8238e" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.239232 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1130beea1319e8fa0fc79c765ffa78d587a366351938b179f4cc04bd68f8238e"} err="failed to get container status \"1130beea1319e8fa0fc79c765ffa78d587a366351938b179f4cc04bd68f8238e\": rpc error: code = NotFound desc = could not find container \"1130beea1319e8fa0fc79c765ffa78d587a366351938b179f4cc04bd68f8238e\": container with ID starting with 1130beea1319e8fa0fc79c765ffa78d587a366351938b179f4cc04bd68f8238e not found: ID does not exist" Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.506181 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n"] Oct 01 16:11:44 crc kubenswrapper[4726]: I1001 16:11:44.579697 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-775b8bd7f8-45s27"] Oct 01 16:11:44 crc kubenswrapper[4726]: W1001 16:11:44.586632 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bcab5a3_6477_47d4_a954_de6f55771a61.slice/crio-641d0a957f1f67d8261cd3f3d5a1b294ef3d7a67e87f52df98e97eb90bb91cd7 WatchSource:0}: Error finding container 641d0a957f1f67d8261cd3f3d5a1b294ef3d7a67e87f52df98e97eb90bb91cd7: Status 404 returned error can't find the container with id 641d0a957f1f67d8261cd3f3d5a1b294ef3d7a67e87f52df98e97eb90bb91cd7 Oct 01 16:11:45 crc kubenswrapper[4726]: I1001 16:11:45.165185 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" event={"ID":"9bcab5a3-6477-47d4-a954-de6f55771a61","Type":"ContainerStarted","Data":"be00354822142f221b0c3f1cc49a62fe596ebfa761da9d67d8d1653af81275f1"} Oct 01 16:11:45 crc kubenswrapper[4726]: I1001 16:11:45.165542 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:45 crc kubenswrapper[4726]: I1001 16:11:45.165556 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" event={"ID":"9bcab5a3-6477-47d4-a954-de6f55771a61","Type":"ContainerStarted","Data":"641d0a957f1f67d8261cd3f3d5a1b294ef3d7a67e87f52df98e97eb90bb91cd7"} Oct 01 16:11:45 crc kubenswrapper[4726]: I1001 16:11:45.166772 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" event={"ID":"0438d3a0-491c-40b3-b596-1940a287b882","Type":"ContainerStarted","Data":"6941d09199869da8da12100bf53abc9c7ee31fc2f95033781169d4e5bd842bbd"} Oct 01 16:11:45 crc kubenswrapper[4726]: I1001 16:11:45.166802 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" event={"ID":"0438d3a0-491c-40b3-b596-1940a287b882","Type":"ContainerStarted","Data":"db47df0230f6193c63ab181e145f2f1f3e66d7eb4c4a0985ef47be55c385eeb2"} Oct 01 16:11:45 crc kubenswrapper[4726]: I1001 16:11:45.167288 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:45 crc kubenswrapper[4726]: I1001 16:11:45.170293 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" Oct 01 16:11:45 crc kubenswrapper[4726]: I1001 16:11:45.172176 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" Oct 01 16:11:45 crc kubenswrapper[4726]: I1001 16:11:45.179268 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-775b8bd7f8-45s27" podStartSLOduration=2.1792499579999998 podStartE2EDuration="2.179249958s" podCreationTimestamp="2025-10-01 16:11:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:11:45.179169175 +0000 UTC m=+758.080721752" watchObservedRunningTime="2025-10-01 16:11:45.179249958 +0000 UTC m=+758.080802535" Oct 01 16:11:45 crc kubenswrapper[4726]: I1001 16:11:45.201193 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-74dbd8bd9b-jz74n" podStartSLOduration=2.201175406 podStartE2EDuration="2.201175406s" podCreationTimestamp="2025-10-01 16:11:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:11:45.1975995 +0000 UTC m=+758.099152067" watchObservedRunningTime="2025-10-01 16:11:45.201175406 +0000 UTC m=+758.102727983" Oct 01 16:11:45 crc kubenswrapper[4726]: I1001 16:11:45.817486 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a8d300c-d03d-4927-b9ba-c17f692a9db4" path="/var/lib/kubelet/pods/1a8d300c-d03d-4927-b9ba-c17f692a9db4/volumes" Oct 01 16:11:45 crc kubenswrapper[4726]: I1001 16:11:45.817996 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73789e0a-a83d-4470-9f4c-137b3c317b85" path="/var/lib/kubelet/pods/73789e0a-a83d-4470-9f4c-137b3c317b85/volumes" Oct 01 16:11:47 crc kubenswrapper[4726]: I1001 16:11:47.495410 4726 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 16:11:48 crc kubenswrapper[4726]: I1001 16:11:48.186093 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" event={"ID":"daa5265d-09f5-4a83-96a0-415448d42bcd","Type":"ContainerStarted","Data":"91ef4d875b12d908c04b4179a391a99b64695c29ccb91c2cf6cfdb806b9e00b0"} Oct 01 16:11:48 crc kubenswrapper[4726]: I1001 16:11:48.186367 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:11:48 crc kubenswrapper[4726]: I1001 16:11:48.187665 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" event={"ID":"94e8d60f-8f3a-448a-9096-163826d5ad92","Type":"ContainerStarted","Data":"f63336a280fb49c32c3b420b5e53614ecd913f739bb694a58658e325e6691317"} Oct 01 16:11:48 crc kubenswrapper[4726]: I1001 16:11:48.188215 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:11:48 crc kubenswrapper[4726]: I1001 16:11:48.201779 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" podStartSLOduration=1.920499288 podStartE2EDuration="7.201765645s" podCreationTimestamp="2025-10-01 16:11:41 +0000 UTC" firstStartedPulling="2025-10-01 16:11:42.512196319 +0000 UTC m=+755.413748896" lastFinishedPulling="2025-10-01 16:11:47.793462686 +0000 UTC m=+760.695015253" observedRunningTime="2025-10-01 16:11:48.200073625 +0000 UTC m=+761.101626202" watchObservedRunningTime="2025-10-01 16:11:48.201765645 +0000 UTC m=+761.103318222" Oct 01 16:11:48 crc kubenswrapper[4726]: I1001 16:11:48.220420 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" podStartSLOduration=1.5925333739999998 podStartE2EDuration="7.220403816s" podCreationTimestamp="2025-10-01 16:11:41 +0000 UTC" firstStartedPulling="2025-10-01 16:11:42.144332325 +0000 UTC m=+755.045884892" lastFinishedPulling="2025-10-01 16:11:47.772202747 +0000 UTC m=+760.673755334" observedRunningTime="2025-10-01 16:11:48.217536312 +0000 UTC m=+761.119088949" watchObservedRunningTime="2025-10-01 16:11:48.220403816 +0000 UTC m=+761.121956393" Oct 01 16:11:53 crc kubenswrapper[4726]: I1001 16:11:53.413683 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:11:53 crc kubenswrapper[4726]: I1001 16:11:53.414507 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:12:02 crc kubenswrapper[4726]: I1001 16:12:02.241021 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:12:21 crc kubenswrapper[4726]: I1001 16:12:21.265212 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9n9c8"] Oct 01 16:12:21 crc kubenswrapper[4726]: I1001 16:12:21.269619 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:21 crc kubenswrapper[4726]: I1001 16:12:21.296433 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9n9c8"] Oct 01 16:12:21 crc kubenswrapper[4726]: I1001 16:12:21.389831 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/863e916f-37fd-48a7-af91-0db2e32d4100-utilities\") pod \"redhat-operators-9n9c8\" (UID: \"863e916f-37fd-48a7-af91-0db2e32d4100\") " pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:21 crc kubenswrapper[4726]: I1001 16:12:21.389883 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/863e916f-37fd-48a7-af91-0db2e32d4100-catalog-content\") pod \"redhat-operators-9n9c8\" (UID: \"863e916f-37fd-48a7-af91-0db2e32d4100\") " pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:21 crc kubenswrapper[4726]: I1001 16:12:21.389952 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x55n4\" (UniqueName: \"kubernetes.io/projected/863e916f-37fd-48a7-af91-0db2e32d4100-kube-api-access-x55n4\") pod \"redhat-operators-9n9c8\" (UID: \"863e916f-37fd-48a7-af91-0db2e32d4100\") " pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:21 crc kubenswrapper[4726]: I1001 16:12:21.491120 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/863e916f-37fd-48a7-af91-0db2e32d4100-utilities\") pod \"redhat-operators-9n9c8\" (UID: \"863e916f-37fd-48a7-af91-0db2e32d4100\") " pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:21 crc kubenswrapper[4726]: I1001 16:12:21.491197 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/863e916f-37fd-48a7-af91-0db2e32d4100-catalog-content\") pod \"redhat-operators-9n9c8\" (UID: \"863e916f-37fd-48a7-af91-0db2e32d4100\") " pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:21 crc kubenswrapper[4726]: I1001 16:12:21.491288 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x55n4\" (UniqueName: \"kubernetes.io/projected/863e916f-37fd-48a7-af91-0db2e32d4100-kube-api-access-x55n4\") pod \"redhat-operators-9n9c8\" (UID: \"863e916f-37fd-48a7-af91-0db2e32d4100\") " pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:21 crc kubenswrapper[4726]: I1001 16:12:21.491671 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/863e916f-37fd-48a7-af91-0db2e32d4100-utilities\") pod \"redhat-operators-9n9c8\" (UID: \"863e916f-37fd-48a7-af91-0db2e32d4100\") " pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:21 crc kubenswrapper[4726]: I1001 16:12:21.491877 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/863e916f-37fd-48a7-af91-0db2e32d4100-catalog-content\") pod \"redhat-operators-9n9c8\" (UID: \"863e916f-37fd-48a7-af91-0db2e32d4100\") " pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:21 crc kubenswrapper[4726]: I1001 16:12:21.525563 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x55n4\" (UniqueName: \"kubernetes.io/projected/863e916f-37fd-48a7-af91-0db2e32d4100-kube-api-access-x55n4\") pod \"redhat-operators-9n9c8\" (UID: \"863e916f-37fd-48a7-af91-0db2e32d4100\") " pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:21 crc kubenswrapper[4726]: I1001 16:12:21.620711 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:21 crc kubenswrapper[4726]: I1001 16:12:21.858573 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.064492 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9n9c8"] Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.403636 4726 generic.go:334] "Generic (PLEG): container finished" podID="863e916f-37fd-48a7-af91-0db2e32d4100" containerID="6bcf4e3faa716c0dd93a794dc0e15258a0e81695f4d44216911f642f395f5afe" exitCode=0 Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.403818 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n9c8" event={"ID":"863e916f-37fd-48a7-af91-0db2e32d4100","Type":"ContainerDied","Data":"6bcf4e3faa716c0dd93a794dc0e15258a0e81695f4d44216911f642f395f5afe"} Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.403892 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n9c8" event={"ID":"863e916f-37fd-48a7-af91-0db2e32d4100","Type":"ContainerStarted","Data":"45a8a4c36ca9a4c4de9317014bbd2122c0cc8c21a470bf5ae74da8b0e1519729"} Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.554351 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-r4djs"] Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.557275 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.557446 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7"] Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.558617 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.559892 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.561557 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.564478 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-gnjxn" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.565301 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.574001 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7"] Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.627900 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-9frd4"] Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.628856 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9frd4" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.630332 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.630559 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.630691 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.631138 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-z5rwg" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.647431 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5d688f5ffc-gr6kx"] Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.648361 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.660900 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.662248 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-gr6kx"] Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.704366 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-reloader\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.704422 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpzp4\" (UniqueName: \"kubernetes.io/projected/0edf23ef-c2c4-4442-89ab-e736c6b989b5-kube-api-access-gpzp4\") pod \"frr-k8s-webhook-server-5478bdb765-cpqd7\" (UID: \"0edf23ef-c2c4-4442-89ab-e736c6b989b5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.704448 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcgnw\" (UniqueName: \"kubernetes.io/projected/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-kube-api-access-lcgnw\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.704473 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-metrics\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.704500 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0edf23ef-c2c4-4442-89ab-e736c6b989b5-cert\") pod \"frr-k8s-webhook-server-5478bdb765-cpqd7\" (UID: \"0edf23ef-c2c4-4442-89ab-e736c6b989b5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.704518 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-startup\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.704535 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-metrics-certs\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.704567 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-conf\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.704586 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-sockets\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806254 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-metrics-certs\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806317 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-conf\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806344 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/08f6790e-97a0-48c4-815d-cb6d57fa935c-metallb-excludel2\") pod \"speaker-9frd4\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " pod="metallb-system/speaker-9frd4" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806363 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-sockets\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806384 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj5dv\" (UniqueName: \"kubernetes.io/projected/152cfa40-5a89-4715-825d-648e9f1c3d2f-kube-api-access-rj5dv\") pod \"controller-5d688f5ffc-gr6kx\" (UID: \"152cfa40-5a89-4715-825d-648e9f1c3d2f\") " pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806404 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-reloader\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806429 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpzp4\" (UniqueName: \"kubernetes.io/projected/0edf23ef-c2c4-4442-89ab-e736c6b989b5-kube-api-access-gpzp4\") pod \"frr-k8s-webhook-server-5478bdb765-cpqd7\" (UID: \"0edf23ef-c2c4-4442-89ab-e736c6b989b5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806448 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g77jf\" (UniqueName: \"kubernetes.io/projected/08f6790e-97a0-48c4-815d-cb6d57fa935c-kube-api-access-g77jf\") pod \"speaker-9frd4\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " pod="metallb-system/speaker-9frd4" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806468 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcgnw\" (UniqueName: \"kubernetes.io/projected/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-kube-api-access-lcgnw\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806482 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/152cfa40-5a89-4715-825d-648e9f1c3d2f-cert\") pod \"controller-5d688f5ffc-gr6kx\" (UID: \"152cfa40-5a89-4715-825d-648e9f1c3d2f\") " pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806527 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-metrics\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806542 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-memberlist\") pod \"speaker-9frd4\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " pod="metallb-system/speaker-9frd4" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806563 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/152cfa40-5a89-4715-825d-648e9f1c3d2f-metrics-certs\") pod \"controller-5d688f5ffc-gr6kx\" (UID: \"152cfa40-5a89-4715-825d-648e9f1c3d2f\") " pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806580 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-metrics-certs\") pod \"speaker-9frd4\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " pod="metallb-system/speaker-9frd4" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806598 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0edf23ef-c2c4-4442-89ab-e736c6b989b5-cert\") pod \"frr-k8s-webhook-server-5478bdb765-cpqd7\" (UID: \"0edf23ef-c2c4-4442-89ab-e736c6b989b5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.806615 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-startup\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.807265 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-conf\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.807343 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-reloader\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.807533 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-sockets\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.807593 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-metrics\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.807646 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-startup\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.812797 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-metrics-certs\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.813447 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0edf23ef-c2c4-4442-89ab-e736c6b989b5-cert\") pod \"frr-k8s-webhook-server-5478bdb765-cpqd7\" (UID: \"0edf23ef-c2c4-4442-89ab-e736c6b989b5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.821296 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcgnw\" (UniqueName: \"kubernetes.io/projected/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-kube-api-access-lcgnw\") pod \"frr-k8s-r4djs\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.826225 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpzp4\" (UniqueName: \"kubernetes.io/projected/0edf23ef-c2c4-4442-89ab-e736c6b989b5-kube-api-access-gpzp4\") pod \"frr-k8s-webhook-server-5478bdb765-cpqd7\" (UID: \"0edf23ef-c2c4-4442-89ab-e736c6b989b5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.874425 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.881819 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.907367 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj5dv\" (UniqueName: \"kubernetes.io/projected/152cfa40-5a89-4715-825d-648e9f1c3d2f-kube-api-access-rj5dv\") pod \"controller-5d688f5ffc-gr6kx\" (UID: \"152cfa40-5a89-4715-825d-648e9f1c3d2f\") " pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.907438 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g77jf\" (UniqueName: \"kubernetes.io/projected/08f6790e-97a0-48c4-815d-cb6d57fa935c-kube-api-access-g77jf\") pod \"speaker-9frd4\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " pod="metallb-system/speaker-9frd4" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.907460 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/152cfa40-5a89-4715-825d-648e9f1c3d2f-cert\") pod \"controller-5d688f5ffc-gr6kx\" (UID: \"152cfa40-5a89-4715-825d-648e9f1c3d2f\") " pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.907486 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-memberlist\") pod \"speaker-9frd4\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " pod="metallb-system/speaker-9frd4" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.907506 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/152cfa40-5a89-4715-825d-648e9f1c3d2f-metrics-certs\") pod \"controller-5d688f5ffc-gr6kx\" (UID: \"152cfa40-5a89-4715-825d-648e9f1c3d2f\") " pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.907521 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-metrics-certs\") pod \"speaker-9frd4\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " pod="metallb-system/speaker-9frd4" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.907574 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/08f6790e-97a0-48c4-815d-cb6d57fa935c-metallb-excludel2\") pod \"speaker-9frd4\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " pod="metallb-system/speaker-9frd4" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.908334 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/08f6790e-97a0-48c4-815d-cb6d57fa935c-metallb-excludel2\") pod \"speaker-9frd4\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " pod="metallb-system/speaker-9frd4" Oct 01 16:12:22 crc kubenswrapper[4726]: E1001 16:12:22.908466 4726 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 01 16:12:22 crc kubenswrapper[4726]: E1001 16:12:22.908560 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-memberlist podName:08f6790e-97a0-48c4-815d-cb6d57fa935c nodeName:}" failed. No retries permitted until 2025-10-01 16:12:23.408532762 +0000 UTC m=+796.310085389 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-memberlist") pod "speaker-9frd4" (UID: "08f6790e-97a0-48c4-815d-cb6d57fa935c") : secret "metallb-memberlist" not found Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.911323 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.913396 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/152cfa40-5a89-4715-825d-648e9f1c3d2f-metrics-certs\") pod \"controller-5d688f5ffc-gr6kx\" (UID: \"152cfa40-5a89-4715-825d-648e9f1c3d2f\") " pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.913397 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-metrics-certs\") pod \"speaker-9frd4\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " pod="metallb-system/speaker-9frd4" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.921432 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/152cfa40-5a89-4715-825d-648e9f1c3d2f-cert\") pod \"controller-5d688f5ffc-gr6kx\" (UID: \"152cfa40-5a89-4715-825d-648e9f1c3d2f\") " pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.924636 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g77jf\" (UniqueName: \"kubernetes.io/projected/08f6790e-97a0-48c4-815d-cb6d57fa935c-kube-api-access-g77jf\") pod \"speaker-9frd4\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " pod="metallb-system/speaker-9frd4" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.928449 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj5dv\" (UniqueName: \"kubernetes.io/projected/152cfa40-5a89-4715-825d-648e9f1c3d2f-kube-api-access-rj5dv\") pod \"controller-5d688f5ffc-gr6kx\" (UID: \"152cfa40-5a89-4715-825d-648e9f1c3d2f\") " pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:12:22 crc kubenswrapper[4726]: I1001 16:12:22.963324 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:12:23 crc kubenswrapper[4726]: I1001 16:12:23.294451 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7"] Oct 01 16:12:23 crc kubenswrapper[4726]: W1001 16:12:23.303877 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0edf23ef_c2c4_4442_89ab_e736c6b989b5.slice/crio-35e89045fafb920b2f096f35706772ba4511889304f17426a25b18009cd4c2dc WatchSource:0}: Error finding container 35e89045fafb920b2f096f35706772ba4511889304f17426a25b18009cd4c2dc: Status 404 returned error can't find the container with id 35e89045fafb920b2f096f35706772ba4511889304f17426a25b18009cd4c2dc Oct 01 16:12:23 crc kubenswrapper[4726]: I1001 16:12:23.412918 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-memberlist\") pod \"speaker-9frd4\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " pod="metallb-system/speaker-9frd4" Oct 01 16:12:23 crc kubenswrapper[4726]: E1001 16:12:23.413192 4726 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 01 16:12:23 crc kubenswrapper[4726]: I1001 16:12:23.413280 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:12:23 crc kubenswrapper[4726]: E1001 16:12:23.413290 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-memberlist podName:08f6790e-97a0-48c4-815d-cb6d57fa935c nodeName:}" failed. No retries permitted until 2025-10-01 16:12:24.413260835 +0000 UTC m=+797.314813452 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-memberlist") pod "speaker-9frd4" (UID: "08f6790e-97a0-48c4-815d-cb6d57fa935c") : secret "metallb-memberlist" not found Oct 01 16:12:23 crc kubenswrapper[4726]: I1001 16:12:23.413314 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:12:23 crc kubenswrapper[4726]: I1001 16:12:23.413398 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 16:12:23 crc kubenswrapper[4726]: I1001 16:12:23.413801 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" event={"ID":"0edf23ef-c2c4-4442-89ab-e736c6b989b5","Type":"ContainerStarted","Data":"35e89045fafb920b2f096f35706772ba4511889304f17426a25b18009cd4c2dc"} Oct 01 16:12:23 crc kubenswrapper[4726]: I1001 16:12:23.413841 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"843d4fde9abf12c4ceb01029d43ed7c5445854afa4d6b76b3edc8ef114cd52e4"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:12:23 crc kubenswrapper[4726]: I1001 16:12:23.413884 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://843d4fde9abf12c4ceb01029d43ed7c5445854afa4d6b76b3edc8ef114cd52e4" gracePeriod=600 Oct 01 16:12:23 crc kubenswrapper[4726]: I1001 16:12:23.414837 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r4djs" event={"ID":"f08bbaf6-372d-43ab-aa18-d6870e95fe4c","Type":"ContainerStarted","Data":"6c19a6336e132b681c509e5f0efcbfb8968d35617274378deb4b650eb438efc7"} Oct 01 16:12:23 crc kubenswrapper[4726]: I1001 16:12:23.423620 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n9c8" event={"ID":"863e916f-37fd-48a7-af91-0db2e32d4100","Type":"ContainerStarted","Data":"050dfc6b4687400d4a24159fb8095db62cc89058af3daf85fcb38df4e814d08b"} Oct 01 16:12:23 crc kubenswrapper[4726]: I1001 16:12:23.467520 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-gr6kx"] Oct 01 16:12:23 crc kubenswrapper[4726]: I1001 16:12:23.842519 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qkkgq"] Oct 01 16:12:23 crc kubenswrapper[4726]: I1001 16:12:23.844175 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:23 crc kubenswrapper[4726]: I1001 16:12:23.861684 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qkkgq"] Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.021724 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w45w\" (UniqueName: \"kubernetes.io/projected/4de76b68-b2f4-4c39-9855-17916f558c01-kube-api-access-8w45w\") pod \"redhat-marketplace-qkkgq\" (UID: \"4de76b68-b2f4-4c39-9855-17916f558c01\") " pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.022178 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de76b68-b2f4-4c39-9855-17916f558c01-catalog-content\") pod \"redhat-marketplace-qkkgq\" (UID: \"4de76b68-b2f4-4c39-9855-17916f558c01\") " pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.022225 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de76b68-b2f4-4c39-9855-17916f558c01-utilities\") pod \"redhat-marketplace-qkkgq\" (UID: \"4de76b68-b2f4-4c39-9855-17916f558c01\") " pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.123151 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w45w\" (UniqueName: \"kubernetes.io/projected/4de76b68-b2f4-4c39-9855-17916f558c01-kube-api-access-8w45w\") pod \"redhat-marketplace-qkkgq\" (UID: \"4de76b68-b2f4-4c39-9855-17916f558c01\") " pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.123496 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de76b68-b2f4-4c39-9855-17916f558c01-catalog-content\") pod \"redhat-marketplace-qkkgq\" (UID: \"4de76b68-b2f4-4c39-9855-17916f558c01\") " pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.123620 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de76b68-b2f4-4c39-9855-17916f558c01-utilities\") pod \"redhat-marketplace-qkkgq\" (UID: \"4de76b68-b2f4-4c39-9855-17916f558c01\") " pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.124016 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de76b68-b2f4-4c39-9855-17916f558c01-catalog-content\") pod \"redhat-marketplace-qkkgq\" (UID: \"4de76b68-b2f4-4c39-9855-17916f558c01\") " pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.124023 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de76b68-b2f4-4c39-9855-17916f558c01-utilities\") pod \"redhat-marketplace-qkkgq\" (UID: \"4de76b68-b2f4-4c39-9855-17916f558c01\") " pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.149179 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w45w\" (UniqueName: \"kubernetes.io/projected/4de76b68-b2f4-4c39-9855-17916f558c01-kube-api-access-8w45w\") pod \"redhat-marketplace-qkkgq\" (UID: \"4de76b68-b2f4-4c39-9855-17916f558c01\") " pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.209080 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.425860 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-memberlist\") pod \"speaker-9frd4\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " pod="metallb-system/speaker-9frd4" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.429697 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-memberlist\") pod \"speaker-9frd4\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " pod="metallb-system/speaker-9frd4" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.431657 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-gr6kx" event={"ID":"152cfa40-5a89-4715-825d-648e9f1c3d2f","Type":"ContainerStarted","Data":"5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73"} Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.431702 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-gr6kx" event={"ID":"152cfa40-5a89-4715-825d-648e9f1c3d2f","Type":"ContainerStarted","Data":"359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99"} Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.431713 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-gr6kx" event={"ID":"152cfa40-5a89-4715-825d-648e9f1c3d2f","Type":"ContainerStarted","Data":"f875bf1c322f210ac5e9d71b7ad7267ccf64659de749f1f43fb642a4b63e6910"} Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.431777 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.436020 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="843d4fde9abf12c4ceb01029d43ed7c5445854afa4d6b76b3edc8ef114cd52e4" exitCode=0 Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.436111 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"843d4fde9abf12c4ceb01029d43ed7c5445854afa4d6b76b3edc8ef114cd52e4"} Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.436139 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"3fcc119df2ac4146cc2bd5a3dfb304e41077427fd468db2dfae53b91e3c841a2"} Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.436160 4726 scope.go:117] "RemoveContainer" containerID="d8eaf0f51eb5e70b61a0a405367175f41860b9a554b95b023ac96b6912d9a4f7" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.438918 4726 generic.go:334] "Generic (PLEG): container finished" podID="863e916f-37fd-48a7-af91-0db2e32d4100" containerID="050dfc6b4687400d4a24159fb8095db62cc89058af3daf85fcb38df4e814d08b" exitCode=0 Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.438960 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n9c8" event={"ID":"863e916f-37fd-48a7-af91-0db2e32d4100","Type":"ContainerDied","Data":"050dfc6b4687400d4a24159fb8095db62cc89058af3daf85fcb38df4e814d08b"} Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.449650 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5d688f5ffc-gr6kx" podStartSLOduration=2.449626902 podStartE2EDuration="2.449626902s" podCreationTimestamp="2025-10-01 16:12:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:12:24.445731337 +0000 UTC m=+797.347283914" watchObservedRunningTime="2025-10-01 16:12:24.449626902 +0000 UTC m=+797.351179479" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.450784 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9frd4" Oct 01 16:12:24 crc kubenswrapper[4726]: I1001 16:12:24.620474 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qkkgq"] Oct 01 16:12:25 crc kubenswrapper[4726]: I1001 16:12:25.449860 4726 generic.go:334] "Generic (PLEG): container finished" podID="4de76b68-b2f4-4c39-9855-17916f558c01" containerID="ec70db184f06b986089492b586eefb24d879f4536897834cff9b0ae87e544478" exitCode=0 Oct 01 16:12:25 crc kubenswrapper[4726]: I1001 16:12:25.449914 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qkkgq" event={"ID":"4de76b68-b2f4-4c39-9855-17916f558c01","Type":"ContainerDied","Data":"ec70db184f06b986089492b586eefb24d879f4536897834cff9b0ae87e544478"} Oct 01 16:12:25 crc kubenswrapper[4726]: I1001 16:12:25.450355 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qkkgq" event={"ID":"4de76b68-b2f4-4c39-9855-17916f558c01","Type":"ContainerStarted","Data":"f5cbdc2572c6ea76b20bbd3bb0a79b4da6e13536c404d7220eaaa37a2798ef36"} Oct 01 16:12:25 crc kubenswrapper[4726]: I1001 16:12:25.465093 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9frd4" event={"ID":"08f6790e-97a0-48c4-815d-cb6d57fa935c","Type":"ContainerStarted","Data":"f0050ea6732902ce8c291ef445507eee0396924e50670b6996c61d4eaa7dc647"} Oct 01 16:12:25 crc kubenswrapper[4726]: I1001 16:12:25.465139 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9frd4" event={"ID":"08f6790e-97a0-48c4-815d-cb6d57fa935c","Type":"ContainerStarted","Data":"e7934c6514acc3b6103da5da3fe1f2fcb7d1b84b3ce5f8e705b31de97f4d0620"} Oct 01 16:12:25 crc kubenswrapper[4726]: I1001 16:12:25.465153 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9frd4" event={"ID":"08f6790e-97a0-48c4-815d-cb6d57fa935c","Type":"ContainerStarted","Data":"903b4a502480eed2191eb16b718eecd70a566d307aa52b535d309c9a11f0d628"} Oct 01 16:12:25 crc kubenswrapper[4726]: I1001 16:12:25.465323 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-9frd4" Oct 01 16:12:25 crc kubenswrapper[4726]: I1001 16:12:25.468230 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n9c8" event={"ID":"863e916f-37fd-48a7-af91-0db2e32d4100","Type":"ContainerStarted","Data":"3b762a7e6162df0f503786bc231ac696b8d5a5599f27ae0cb0daa2db46cc5929"} Oct 01 16:12:25 crc kubenswrapper[4726]: I1001 16:12:25.500436 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-9frd4" podStartSLOduration=3.500416652 podStartE2EDuration="3.500416652s" podCreationTimestamp="2025-10-01 16:12:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:12:25.497022882 +0000 UTC m=+798.398575469" watchObservedRunningTime="2025-10-01 16:12:25.500416652 +0000 UTC m=+798.401969229" Oct 01 16:12:25 crc kubenswrapper[4726]: I1001 16:12:25.519935 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9n9c8" podStartSLOduration=2.020877463 podStartE2EDuration="4.519920225s" podCreationTimestamp="2025-10-01 16:12:21 +0000 UTC" firstStartedPulling="2025-10-01 16:12:22.405178839 +0000 UTC m=+795.306731426" lastFinishedPulling="2025-10-01 16:12:24.904221611 +0000 UTC m=+797.805774188" observedRunningTime="2025-10-01 16:12:25.517163294 +0000 UTC m=+798.418715881" watchObservedRunningTime="2025-10-01 16:12:25.519920225 +0000 UTC m=+798.421472802" Oct 01 16:12:26 crc kubenswrapper[4726]: I1001 16:12:26.476468 4726 generic.go:334] "Generic (PLEG): container finished" podID="4de76b68-b2f4-4c39-9855-17916f558c01" containerID="f3b043cc6d1e865a2dddc84d8290338408e7bab35805b692533402ac49184d52" exitCode=0 Oct 01 16:12:26 crc kubenswrapper[4726]: I1001 16:12:26.478569 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qkkgq" event={"ID":"4de76b68-b2f4-4c39-9855-17916f558c01","Type":"ContainerDied","Data":"f3b043cc6d1e865a2dddc84d8290338408e7bab35805b692533402ac49184d52"} Oct 01 16:12:27 crc kubenswrapper[4726]: I1001 16:12:27.488882 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qkkgq" event={"ID":"4de76b68-b2f4-4c39-9855-17916f558c01","Type":"ContainerStarted","Data":"30d7e944b97e7d6c7de316d849292ed1c8994a300f47b12f43782b3bcb6aeb8c"} Oct 01 16:12:27 crc kubenswrapper[4726]: I1001 16:12:27.508039 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qkkgq" podStartSLOduration=2.983624593 podStartE2EDuration="4.508023813s" podCreationTimestamp="2025-10-01 16:12:23 +0000 UTC" firstStartedPulling="2025-10-01 16:12:25.454595995 +0000 UTC m=+798.356148572" lastFinishedPulling="2025-10-01 16:12:26.978995215 +0000 UTC m=+799.880547792" observedRunningTime="2025-10-01 16:12:27.504797758 +0000 UTC m=+800.406350345" watchObservedRunningTime="2025-10-01 16:12:27.508023813 +0000 UTC m=+800.409576390" Oct 01 16:12:31 crc kubenswrapper[4726]: I1001 16:12:31.523106 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" event={"ID":"0edf23ef-c2c4-4442-89ab-e736c6b989b5","Type":"ContainerStarted","Data":"c82a365f75111727835d96f54c55d87e42c823102a75b16d08f3f478beee622d"} Oct 01 16:12:31 crc kubenswrapper[4726]: I1001 16:12:31.523590 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" Oct 01 16:12:31 crc kubenswrapper[4726]: I1001 16:12:31.525113 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r4djs" event={"ID":"f08bbaf6-372d-43ab-aa18-d6870e95fe4c","Type":"ContainerDied","Data":"d4481c40814f006b8e43ba761624c659b826b61501a2ac3cae57677c18dc7015"} Oct 01 16:12:31 crc kubenswrapper[4726]: I1001 16:12:31.524957 4726 generic.go:334] "Generic (PLEG): container finished" podID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerID="d4481c40814f006b8e43ba761624c659b826b61501a2ac3cae57677c18dc7015" exitCode=0 Oct 01 16:12:31 crc kubenswrapper[4726]: I1001 16:12:31.547653 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" podStartSLOduration=1.859881949 podStartE2EDuration="9.54762759s" podCreationTimestamp="2025-10-01 16:12:22 +0000 UTC" firstStartedPulling="2025-10-01 16:12:23.306106186 +0000 UTC m=+796.207658763" lastFinishedPulling="2025-10-01 16:12:30.993851787 +0000 UTC m=+803.895404404" observedRunningTime="2025-10-01 16:12:31.545481487 +0000 UTC m=+804.447034054" watchObservedRunningTime="2025-10-01 16:12:31.54762759 +0000 UTC m=+804.449180207" Oct 01 16:12:31 crc kubenswrapper[4726]: I1001 16:12:31.621576 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:31 crc kubenswrapper[4726]: I1001 16:12:31.621714 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:31 crc kubenswrapper[4726]: I1001 16:12:31.716816 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:32 crc kubenswrapper[4726]: I1001 16:12:32.536756 4726 generic.go:334] "Generic (PLEG): container finished" podID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerID="1a181e670ed2692add3ac9550fd2f23b318e9905f01e3cc543a3dfe70da5d43a" exitCode=0 Oct 01 16:12:32 crc kubenswrapper[4726]: I1001 16:12:32.536923 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r4djs" event={"ID":"f08bbaf6-372d-43ab-aa18-d6870e95fe4c","Type":"ContainerDied","Data":"1a181e670ed2692add3ac9550fd2f23b318e9905f01e3cc543a3dfe70da5d43a"} Oct 01 16:12:32 crc kubenswrapper[4726]: I1001 16:12:32.597810 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:32 crc kubenswrapper[4726]: I1001 16:12:32.644266 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9n9c8"] Oct 01 16:12:34 crc kubenswrapper[4726]: I1001 16:12:34.210239 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:34 crc kubenswrapper[4726]: I1001 16:12:34.210858 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:34 crc kubenswrapper[4726]: I1001 16:12:34.277004 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:34 crc kubenswrapper[4726]: I1001 16:12:34.456938 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-9frd4" Oct 01 16:12:34 crc kubenswrapper[4726]: I1001 16:12:34.554966 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9n9c8" podUID="863e916f-37fd-48a7-af91-0db2e32d4100" containerName="registry-server" containerID="cri-o://3b762a7e6162df0f503786bc231ac696b8d5a5599f27ae0cb0daa2db46cc5929" gracePeriod=2 Oct 01 16:12:34 crc kubenswrapper[4726]: I1001 16:12:34.621400 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:35 crc kubenswrapper[4726]: I1001 16:12:35.348043 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qkkgq"] Oct 01 16:12:35 crc kubenswrapper[4726]: I1001 16:12:35.568676 4726 generic.go:334] "Generic (PLEG): container finished" podID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerID="6c16210952121838cd718e7d1308ee55d2666e7f4be17b4431ce6f04bbf7195e" exitCode=0 Oct 01 16:12:35 crc kubenswrapper[4726]: I1001 16:12:35.568779 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r4djs" event={"ID":"f08bbaf6-372d-43ab-aa18-d6870e95fe4c","Type":"ContainerDied","Data":"6c16210952121838cd718e7d1308ee55d2666e7f4be17b4431ce6f04bbf7195e"} Oct 01 16:12:35 crc kubenswrapper[4726]: I1001 16:12:35.579689 4726 generic.go:334] "Generic (PLEG): container finished" podID="863e916f-37fd-48a7-af91-0db2e32d4100" containerID="3b762a7e6162df0f503786bc231ac696b8d5a5599f27ae0cb0daa2db46cc5929" exitCode=0 Oct 01 16:12:35 crc kubenswrapper[4726]: I1001 16:12:35.579800 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n9c8" event={"ID":"863e916f-37fd-48a7-af91-0db2e32d4100","Type":"ContainerDied","Data":"3b762a7e6162df0f503786bc231ac696b8d5a5599f27ae0cb0daa2db46cc5929"} Oct 01 16:12:35 crc kubenswrapper[4726]: I1001 16:12:35.859557 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:35 crc kubenswrapper[4726]: I1001 16:12:35.983793 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/863e916f-37fd-48a7-af91-0db2e32d4100-catalog-content\") pod \"863e916f-37fd-48a7-af91-0db2e32d4100\" (UID: \"863e916f-37fd-48a7-af91-0db2e32d4100\") " Oct 01 16:12:35 crc kubenswrapper[4726]: I1001 16:12:35.983844 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x55n4\" (UniqueName: \"kubernetes.io/projected/863e916f-37fd-48a7-af91-0db2e32d4100-kube-api-access-x55n4\") pod \"863e916f-37fd-48a7-af91-0db2e32d4100\" (UID: \"863e916f-37fd-48a7-af91-0db2e32d4100\") " Oct 01 16:12:35 crc kubenswrapper[4726]: I1001 16:12:35.983873 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/863e916f-37fd-48a7-af91-0db2e32d4100-utilities\") pod \"863e916f-37fd-48a7-af91-0db2e32d4100\" (UID: \"863e916f-37fd-48a7-af91-0db2e32d4100\") " Oct 01 16:12:35 crc kubenswrapper[4726]: I1001 16:12:35.986213 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/863e916f-37fd-48a7-af91-0db2e32d4100-utilities" (OuterVolumeSpecName: "utilities") pod "863e916f-37fd-48a7-af91-0db2e32d4100" (UID: "863e916f-37fd-48a7-af91-0db2e32d4100"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:12:35 crc kubenswrapper[4726]: I1001 16:12:35.998282 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/863e916f-37fd-48a7-af91-0db2e32d4100-kube-api-access-x55n4" (OuterVolumeSpecName: "kube-api-access-x55n4") pod "863e916f-37fd-48a7-af91-0db2e32d4100" (UID: "863e916f-37fd-48a7-af91-0db2e32d4100"). InnerVolumeSpecName "kube-api-access-x55n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.078208 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/863e916f-37fd-48a7-af91-0db2e32d4100-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "863e916f-37fd-48a7-af91-0db2e32d4100" (UID: "863e916f-37fd-48a7-af91-0db2e32d4100"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.085462 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/863e916f-37fd-48a7-af91-0db2e32d4100-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.085484 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x55n4\" (UniqueName: \"kubernetes.io/projected/863e916f-37fd-48a7-af91-0db2e32d4100-kube-api-access-x55n4\") on node \"crc\" DevicePath \"\"" Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.085494 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/863e916f-37fd-48a7-af91-0db2e32d4100-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.593694 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9n9c8" Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.593694 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n9c8" event={"ID":"863e916f-37fd-48a7-af91-0db2e32d4100","Type":"ContainerDied","Data":"45a8a4c36ca9a4c4de9317014bbd2122c0cc8c21a470bf5ae74da8b0e1519729"} Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.593861 4726 scope.go:117] "RemoveContainer" containerID="3b762a7e6162df0f503786bc231ac696b8d5a5599f27ae0cb0daa2db46cc5929" Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.612626 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r4djs" event={"ID":"f08bbaf6-372d-43ab-aa18-d6870e95fe4c","Type":"ContainerStarted","Data":"52c0e14d3038a74f8536411a7900dff96b2a9ae673439bf230ae87996a239754"} Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.612726 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r4djs" event={"ID":"f08bbaf6-372d-43ab-aa18-d6870e95fe4c","Type":"ContainerStarted","Data":"abd2a7e13f99d37144d0076c20361976ce336847645daf450d24449bbadfe0ab"} Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.612754 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r4djs" event={"ID":"f08bbaf6-372d-43ab-aa18-d6870e95fe4c","Type":"ContainerStarted","Data":"33e1649f8c3dabd843a405caf4386c07f38d0c68d7380a48b0cbc7e7e679c3cc"} Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.612818 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r4djs" event={"ID":"f08bbaf6-372d-43ab-aa18-d6870e95fe4c","Type":"ContainerStarted","Data":"b15081c4adf77afe809fbd87b422a48727f4320cbfa77d910c72a963c0b485f6"} Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.612840 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r4djs" event={"ID":"f08bbaf6-372d-43ab-aa18-d6870e95fe4c","Type":"ContainerStarted","Data":"a1374a43ccb81a9a3f97068c3ca67cf90019219e90d280cff217e2b3c54782ef"} Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.612889 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qkkgq" podUID="4de76b68-b2f4-4c39-9855-17916f558c01" containerName="registry-server" containerID="cri-o://30d7e944b97e7d6c7de316d849292ed1c8994a300f47b12f43782b3bcb6aeb8c" gracePeriod=2 Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.635761 4726 scope.go:117] "RemoveContainer" containerID="050dfc6b4687400d4a24159fb8095db62cc89058af3daf85fcb38df4e814d08b" Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.638251 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9n9c8"] Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.646149 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9n9c8"] Oct 01 16:12:36 crc kubenswrapper[4726]: I1001 16:12:36.738831 4726 scope.go:117] "RemoveContainer" containerID="6bcf4e3faa716c0dd93a794dc0e15258a0e81695f4d44216911f642f395f5afe" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:36.997443 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.099983 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w45w\" (UniqueName: \"kubernetes.io/projected/4de76b68-b2f4-4c39-9855-17916f558c01-kube-api-access-8w45w\") pod \"4de76b68-b2f4-4c39-9855-17916f558c01\" (UID: \"4de76b68-b2f4-4c39-9855-17916f558c01\") " Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.100071 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de76b68-b2f4-4c39-9855-17916f558c01-utilities\") pod \"4de76b68-b2f4-4c39-9855-17916f558c01\" (UID: \"4de76b68-b2f4-4c39-9855-17916f558c01\") " Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.100157 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de76b68-b2f4-4c39-9855-17916f558c01-catalog-content\") pod \"4de76b68-b2f4-4c39-9855-17916f558c01\" (UID: \"4de76b68-b2f4-4c39-9855-17916f558c01\") " Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.101177 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4de76b68-b2f4-4c39-9855-17916f558c01-utilities" (OuterVolumeSpecName: "utilities") pod "4de76b68-b2f4-4c39-9855-17916f558c01" (UID: "4de76b68-b2f4-4c39-9855-17916f558c01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.106229 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4de76b68-b2f4-4c39-9855-17916f558c01-kube-api-access-8w45w" (OuterVolumeSpecName: "kube-api-access-8w45w") pod "4de76b68-b2f4-4c39-9855-17916f558c01" (UID: "4de76b68-b2f4-4c39-9855-17916f558c01"). InnerVolumeSpecName "kube-api-access-8w45w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.112529 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4de76b68-b2f4-4c39-9855-17916f558c01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4de76b68-b2f4-4c39-9855-17916f558c01" (UID: "4de76b68-b2f4-4c39-9855-17916f558c01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.202093 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w45w\" (UniqueName: \"kubernetes.io/projected/4de76b68-b2f4-4c39-9855-17916f558c01-kube-api-access-8w45w\") on node \"crc\" DevicePath \"\"" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.202141 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de76b68-b2f4-4c39-9855-17916f558c01-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.202162 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de76b68-b2f4-4c39-9855-17916f558c01-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.625937 4726 generic.go:334] "Generic (PLEG): container finished" podID="4de76b68-b2f4-4c39-9855-17916f558c01" containerID="30d7e944b97e7d6c7de316d849292ed1c8994a300f47b12f43782b3bcb6aeb8c" exitCode=0 Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.626038 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qkkgq" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.626077 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qkkgq" event={"ID":"4de76b68-b2f4-4c39-9855-17916f558c01","Type":"ContainerDied","Data":"30d7e944b97e7d6c7de316d849292ed1c8994a300f47b12f43782b3bcb6aeb8c"} Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.626691 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qkkgq" event={"ID":"4de76b68-b2f4-4c39-9855-17916f558c01","Type":"ContainerDied","Data":"f5cbdc2572c6ea76b20bbd3bb0a79b4da6e13536c404d7220eaaa37a2798ef36"} Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.626820 4726 scope.go:117] "RemoveContainer" containerID="30d7e944b97e7d6c7de316d849292ed1c8994a300f47b12f43782b3bcb6aeb8c" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.634882 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r4djs" event={"ID":"f08bbaf6-372d-43ab-aa18-d6870e95fe4c","Type":"ContainerStarted","Data":"21e25a53c4de4535261a28d2622d716c55aa1ae4b3e304ee4cf0c9689471cb6c"} Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.635321 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.662334 4726 scope.go:117] "RemoveContainer" containerID="f3b043cc6d1e865a2dddc84d8290338408e7bab35805b692533402ac49184d52" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.696185 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-r4djs" podStartSLOduration=7.843895122 podStartE2EDuration="15.696159477s" podCreationTimestamp="2025-10-01 16:12:22 +0000 UTC" firstStartedPulling="2025-10-01 16:12:23.140560681 +0000 UTC m=+796.042113258" lastFinishedPulling="2025-10-01 16:12:30.992825006 +0000 UTC m=+803.894377613" observedRunningTime="2025-10-01 16:12:37.692806438 +0000 UTC m=+810.594359055" watchObservedRunningTime="2025-10-01 16:12:37.696159477 +0000 UTC m=+810.597712064" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.704839 4726 scope.go:117] "RemoveContainer" containerID="ec70db184f06b986089492b586eefb24d879f4536897834cff9b0ae87e544478" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.713157 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qkkgq"] Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.718970 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qkkgq"] Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.755010 4726 scope.go:117] "RemoveContainer" containerID="30d7e944b97e7d6c7de316d849292ed1c8994a300f47b12f43782b3bcb6aeb8c" Oct 01 16:12:37 crc kubenswrapper[4726]: E1001 16:12:37.755605 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30d7e944b97e7d6c7de316d849292ed1c8994a300f47b12f43782b3bcb6aeb8c\": container with ID starting with 30d7e944b97e7d6c7de316d849292ed1c8994a300f47b12f43782b3bcb6aeb8c not found: ID does not exist" containerID="30d7e944b97e7d6c7de316d849292ed1c8994a300f47b12f43782b3bcb6aeb8c" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.755651 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30d7e944b97e7d6c7de316d849292ed1c8994a300f47b12f43782b3bcb6aeb8c"} err="failed to get container status \"30d7e944b97e7d6c7de316d849292ed1c8994a300f47b12f43782b3bcb6aeb8c\": rpc error: code = NotFound desc = could not find container \"30d7e944b97e7d6c7de316d849292ed1c8994a300f47b12f43782b3bcb6aeb8c\": container with ID starting with 30d7e944b97e7d6c7de316d849292ed1c8994a300f47b12f43782b3bcb6aeb8c not found: ID does not exist" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.755690 4726 scope.go:117] "RemoveContainer" containerID="f3b043cc6d1e865a2dddc84d8290338408e7bab35805b692533402ac49184d52" Oct 01 16:12:37 crc kubenswrapper[4726]: E1001 16:12:37.756142 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3b043cc6d1e865a2dddc84d8290338408e7bab35805b692533402ac49184d52\": container with ID starting with f3b043cc6d1e865a2dddc84d8290338408e7bab35805b692533402ac49184d52 not found: ID does not exist" containerID="f3b043cc6d1e865a2dddc84d8290338408e7bab35805b692533402ac49184d52" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.756183 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3b043cc6d1e865a2dddc84d8290338408e7bab35805b692533402ac49184d52"} err="failed to get container status \"f3b043cc6d1e865a2dddc84d8290338408e7bab35805b692533402ac49184d52\": rpc error: code = NotFound desc = could not find container \"f3b043cc6d1e865a2dddc84d8290338408e7bab35805b692533402ac49184d52\": container with ID starting with f3b043cc6d1e865a2dddc84d8290338408e7bab35805b692533402ac49184d52 not found: ID does not exist" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.756208 4726 scope.go:117] "RemoveContainer" containerID="ec70db184f06b986089492b586eefb24d879f4536897834cff9b0ae87e544478" Oct 01 16:12:37 crc kubenswrapper[4726]: E1001 16:12:37.756550 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec70db184f06b986089492b586eefb24d879f4536897834cff9b0ae87e544478\": container with ID starting with ec70db184f06b986089492b586eefb24d879f4536897834cff9b0ae87e544478 not found: ID does not exist" containerID="ec70db184f06b986089492b586eefb24d879f4536897834cff9b0ae87e544478" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.756588 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec70db184f06b986089492b586eefb24d879f4536897834cff9b0ae87e544478"} err="failed to get container status \"ec70db184f06b986089492b586eefb24d879f4536897834cff9b0ae87e544478\": rpc error: code = NotFound desc = could not find container \"ec70db184f06b986089492b586eefb24d879f4536897834cff9b0ae87e544478\": container with ID starting with ec70db184f06b986089492b586eefb24d879f4536897834cff9b0ae87e544478 not found: ID does not exist" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.818633 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4de76b68-b2f4-4c39-9855-17916f558c01" path="/var/lib/kubelet/pods/4de76b68-b2f4-4c39-9855-17916f558c01/volumes" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.819382 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="863e916f-37fd-48a7-af91-0db2e32d4100" path="/var/lib/kubelet/pods/863e916f-37fd-48a7-af91-0db2e32d4100/volumes" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.875643 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:37 crc kubenswrapper[4726]: I1001 16:12:37.949620 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.565658 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-wh8qd"] Oct 01 16:12:41 crc kubenswrapper[4726]: E1001 16:12:41.567450 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4de76b68-b2f4-4c39-9855-17916f558c01" containerName="extract-utilities" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.567476 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4de76b68-b2f4-4c39-9855-17916f558c01" containerName="extract-utilities" Oct 01 16:12:41 crc kubenswrapper[4726]: E1001 16:12:41.567538 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4de76b68-b2f4-4c39-9855-17916f558c01" containerName="extract-content" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.567637 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4de76b68-b2f4-4c39-9855-17916f558c01" containerName="extract-content" Oct 01 16:12:41 crc kubenswrapper[4726]: E1001 16:12:41.567686 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4de76b68-b2f4-4c39-9855-17916f558c01" containerName="registry-server" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.567707 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4de76b68-b2f4-4c39-9855-17916f558c01" containerName="registry-server" Oct 01 16:12:41 crc kubenswrapper[4726]: E1001 16:12:41.567735 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="863e916f-37fd-48a7-af91-0db2e32d4100" containerName="extract-utilities" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.567752 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="863e916f-37fd-48a7-af91-0db2e32d4100" containerName="extract-utilities" Oct 01 16:12:41 crc kubenswrapper[4726]: E1001 16:12:41.567781 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="863e916f-37fd-48a7-af91-0db2e32d4100" containerName="extract-content" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.567793 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="863e916f-37fd-48a7-af91-0db2e32d4100" containerName="extract-content" Oct 01 16:12:41 crc kubenswrapper[4726]: E1001 16:12:41.567821 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="863e916f-37fd-48a7-af91-0db2e32d4100" containerName="registry-server" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.567833 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="863e916f-37fd-48a7-af91-0db2e32d4100" containerName="registry-server" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.568109 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4de76b68-b2f4-4c39-9855-17916f558c01" containerName="registry-server" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.568140 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="863e916f-37fd-48a7-af91-0db2e32d4100" containerName="registry-server" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.568773 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wh8qd" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.572425 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.578007 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.579997 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-l6ng9" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.580801 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wh8qd"] Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.770087 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9brm6\" (UniqueName: \"kubernetes.io/projected/62b6b99f-a44a-445b-9183-917d39934259-kube-api-access-9brm6\") pod \"openstack-operator-index-wh8qd\" (UID: \"62b6b99f-a44a-445b-9183-917d39934259\") " pod="openstack-operators/openstack-operator-index-wh8qd" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.872329 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9brm6\" (UniqueName: \"kubernetes.io/projected/62b6b99f-a44a-445b-9183-917d39934259-kube-api-access-9brm6\") pod \"openstack-operator-index-wh8qd\" (UID: \"62b6b99f-a44a-445b-9183-917d39934259\") " pod="openstack-operators/openstack-operator-index-wh8qd" Oct 01 16:12:41 crc kubenswrapper[4726]: I1001 16:12:41.906923 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9brm6\" (UniqueName: \"kubernetes.io/projected/62b6b99f-a44a-445b-9183-917d39934259-kube-api-access-9brm6\") pod \"openstack-operator-index-wh8qd\" (UID: \"62b6b99f-a44a-445b-9183-917d39934259\") " pod="openstack-operators/openstack-operator-index-wh8qd" Oct 01 16:12:42 crc kubenswrapper[4726]: I1001 16:12:42.201950 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wh8qd" Oct 01 16:12:42 crc kubenswrapper[4726]: I1001 16:12:42.703879 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wh8qd"] Oct 01 16:12:42 crc kubenswrapper[4726]: I1001 16:12:42.893152 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" Oct 01 16:12:42 crc kubenswrapper[4726]: I1001 16:12:42.970892 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:12:43 crc kubenswrapper[4726]: I1001 16:12:43.684721 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wh8qd" event={"ID":"62b6b99f-a44a-445b-9183-917d39934259","Type":"ContainerStarted","Data":"22c4c7e9f0ee0c10b9b32e2800eca6cff81cac77b83ce61193d90a5903968e11"} Oct 01 16:12:45 crc kubenswrapper[4726]: I1001 16:12:45.702327 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wh8qd" event={"ID":"62b6b99f-a44a-445b-9183-917d39934259","Type":"ContainerStarted","Data":"e22d7c3abfc3cb7f99a85af6dc201e6f9d0823d0008244564d3f0f4df97e79fd"} Oct 01 16:12:45 crc kubenswrapper[4726]: I1001 16:12:45.732852 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-wh8qd" podStartSLOduration=2.078516525 podStartE2EDuration="4.732826351s" podCreationTimestamp="2025-10-01 16:12:41 +0000 UTC" firstStartedPulling="2025-10-01 16:12:42.724236544 +0000 UTC m=+815.625789131" lastFinishedPulling="2025-10-01 16:12:45.37854634 +0000 UTC m=+818.280098957" observedRunningTime="2025-10-01 16:12:45.72667261 +0000 UTC m=+818.628225217" watchObservedRunningTime="2025-10-01 16:12:45.732826351 +0000 UTC m=+818.634378968" Oct 01 16:12:52 crc kubenswrapper[4726]: I1001 16:12:52.203280 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-wh8qd" Oct 01 16:12:52 crc kubenswrapper[4726]: I1001 16:12:52.204032 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-wh8qd" Oct 01 16:12:52 crc kubenswrapper[4726]: I1001 16:12:52.250436 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-wh8qd" Oct 01 16:12:52 crc kubenswrapper[4726]: I1001 16:12:52.796374 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-wh8qd" Oct 01 16:12:52 crc kubenswrapper[4726]: I1001 16:12:52.885685 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-r4djs" Oct 01 16:12:55 crc kubenswrapper[4726]: I1001 16:12:55.817200 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd"] Oct 01 16:12:55 crc kubenswrapper[4726]: I1001 16:12:55.819088 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" Oct 01 16:12:55 crc kubenswrapper[4726]: I1001 16:12:55.824964 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pzrkg" Oct 01 16:12:55 crc kubenswrapper[4726]: I1001 16:12:55.830737 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd"] Oct 01 16:12:56 crc kubenswrapper[4726]: I1001 16:12:56.003827 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-util\") pod \"62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd\" (UID: \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\") " pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" Oct 01 16:12:56 crc kubenswrapper[4726]: I1001 16:12:56.005012 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6psb\" (UniqueName: \"kubernetes.io/projected/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-kube-api-access-n6psb\") pod \"62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd\" (UID: \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\") " pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" Oct 01 16:12:56 crc kubenswrapper[4726]: I1001 16:12:56.005109 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-bundle\") pod \"62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd\" (UID: \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\") " pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" Oct 01 16:12:56 crc kubenswrapper[4726]: I1001 16:12:56.106654 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6psb\" (UniqueName: \"kubernetes.io/projected/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-kube-api-access-n6psb\") pod \"62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd\" (UID: \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\") " pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" Oct 01 16:12:56 crc kubenswrapper[4726]: I1001 16:12:56.106730 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-bundle\") pod \"62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd\" (UID: \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\") " pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" Oct 01 16:12:56 crc kubenswrapper[4726]: I1001 16:12:56.106805 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-util\") pod \"62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd\" (UID: \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\") " pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" Oct 01 16:12:56 crc kubenswrapper[4726]: I1001 16:12:56.107460 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-util\") pod \"62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd\" (UID: \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\") " pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" Oct 01 16:12:56 crc kubenswrapper[4726]: I1001 16:12:56.107597 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-bundle\") pod \"62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd\" (UID: \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\") " pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" Oct 01 16:12:56 crc kubenswrapper[4726]: I1001 16:12:56.134351 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6psb\" (UniqueName: \"kubernetes.io/projected/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-kube-api-access-n6psb\") pod \"62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd\" (UID: \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\") " pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" Oct 01 16:12:56 crc kubenswrapper[4726]: I1001 16:12:56.149702 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" Oct 01 16:12:56 crc kubenswrapper[4726]: I1001 16:12:56.415352 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd"] Oct 01 16:12:56 crc kubenswrapper[4726]: W1001 16:12:56.419921 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6eef09a6_cce6_45d8_ac7e_6b09f74d730a.slice/crio-a2b29750d45ed9ea70914aa49095eee06cc1f6bd726390a8f432434aa3fdca61 WatchSource:0}: Error finding container a2b29750d45ed9ea70914aa49095eee06cc1f6bd726390a8f432434aa3fdca61: Status 404 returned error can't find the container with id a2b29750d45ed9ea70914aa49095eee06cc1f6bd726390a8f432434aa3fdca61 Oct 01 16:12:56 crc kubenswrapper[4726]: I1001 16:12:56.796670 4726 generic.go:334] "Generic (PLEG): container finished" podID="6eef09a6-cce6-45d8-ac7e-6b09f74d730a" containerID="663f669b89b77fbb0c4cd97d1bbb60387cb621031556e8460f9e98789cd3872b" exitCode=0 Oct 01 16:12:56 crc kubenswrapper[4726]: I1001 16:12:56.796752 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" event={"ID":"6eef09a6-cce6-45d8-ac7e-6b09f74d730a","Type":"ContainerDied","Data":"663f669b89b77fbb0c4cd97d1bbb60387cb621031556e8460f9e98789cd3872b"} Oct 01 16:12:56 crc kubenswrapper[4726]: I1001 16:12:56.796847 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" event={"ID":"6eef09a6-cce6-45d8-ac7e-6b09f74d730a","Type":"ContainerStarted","Data":"a2b29750d45ed9ea70914aa49095eee06cc1f6bd726390a8f432434aa3fdca61"} Oct 01 16:12:57 crc kubenswrapper[4726]: I1001 16:12:57.803285 4726 generic.go:334] "Generic (PLEG): container finished" podID="6eef09a6-cce6-45d8-ac7e-6b09f74d730a" containerID="7f92e0173c4f80388d095cef662a85a83273cb1d25bed1babb8ec35cc25d3d18" exitCode=0 Oct 01 16:12:57 crc kubenswrapper[4726]: I1001 16:12:57.803334 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" event={"ID":"6eef09a6-cce6-45d8-ac7e-6b09f74d730a","Type":"ContainerDied","Data":"7f92e0173c4f80388d095cef662a85a83273cb1d25bed1babb8ec35cc25d3d18"} Oct 01 16:12:58 crc kubenswrapper[4726]: I1001 16:12:58.814957 4726 generic.go:334] "Generic (PLEG): container finished" podID="6eef09a6-cce6-45d8-ac7e-6b09f74d730a" containerID="4576496c532bd1c8fa2dc1cb10fcb38c94280a7a3fc472141c27a60db46d0c09" exitCode=0 Oct 01 16:12:58 crc kubenswrapper[4726]: I1001 16:12:58.815031 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" event={"ID":"6eef09a6-cce6-45d8-ac7e-6b09f74d730a","Type":"ContainerDied","Data":"4576496c532bd1c8fa2dc1cb10fcb38c94280a7a3fc472141c27a60db46d0c09"} Oct 01 16:13:00 crc kubenswrapper[4726]: I1001 16:13:00.170113 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" Oct 01 16:13:00 crc kubenswrapper[4726]: I1001 16:13:00.365084 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-bundle\") pod \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\" (UID: \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\") " Oct 01 16:13:00 crc kubenswrapper[4726]: I1001 16:13:00.365207 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-util\") pod \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\" (UID: \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\") " Oct 01 16:13:00 crc kubenswrapper[4726]: I1001 16:13:00.365355 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6psb\" (UniqueName: \"kubernetes.io/projected/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-kube-api-access-n6psb\") pod \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\" (UID: \"6eef09a6-cce6-45d8-ac7e-6b09f74d730a\") " Oct 01 16:13:00 crc kubenswrapper[4726]: I1001 16:13:00.365868 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-bundle" (OuterVolumeSpecName: "bundle") pod "6eef09a6-cce6-45d8-ac7e-6b09f74d730a" (UID: "6eef09a6-cce6-45d8-ac7e-6b09f74d730a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:13:00 crc kubenswrapper[4726]: I1001 16:13:00.370175 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-kube-api-access-n6psb" (OuterVolumeSpecName: "kube-api-access-n6psb") pod "6eef09a6-cce6-45d8-ac7e-6b09f74d730a" (UID: "6eef09a6-cce6-45d8-ac7e-6b09f74d730a"). InnerVolumeSpecName "kube-api-access-n6psb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:13:00 crc kubenswrapper[4726]: I1001 16:13:00.387637 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-util" (OuterVolumeSpecName: "util") pod "6eef09a6-cce6-45d8-ac7e-6b09f74d730a" (UID: "6eef09a6-cce6-45d8-ac7e-6b09f74d730a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:13:00 crc kubenswrapper[4726]: I1001 16:13:00.466327 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6psb\" (UniqueName: \"kubernetes.io/projected/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-kube-api-access-n6psb\") on node \"crc\" DevicePath \"\"" Oct 01 16:13:00 crc kubenswrapper[4726]: I1001 16:13:00.466370 4726 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:13:00 crc kubenswrapper[4726]: I1001 16:13:00.466390 4726 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a6-cce6-45d8-ac7e-6b09f74d730a-util\") on node \"crc\" DevicePath \"\"" Oct 01 16:13:00 crc kubenswrapper[4726]: I1001 16:13:00.835423 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" event={"ID":"6eef09a6-cce6-45d8-ac7e-6b09f74d730a","Type":"ContainerDied","Data":"a2b29750d45ed9ea70914aa49095eee06cc1f6bd726390a8f432434aa3fdca61"} Oct 01 16:13:00 crc kubenswrapper[4726]: I1001 16:13:00.835486 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd" Oct 01 16:13:00 crc kubenswrapper[4726]: I1001 16:13:00.835490 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2b29750d45ed9ea70914aa49095eee06cc1f6bd726390a8f432434aa3fdca61" Oct 01 16:13:05 crc kubenswrapper[4726]: I1001 16:13:05.124906 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6877df559-w8rr9"] Oct 01 16:13:05 crc kubenswrapper[4726]: E1001 16:13:05.125702 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eef09a6-cce6-45d8-ac7e-6b09f74d730a" containerName="extract" Oct 01 16:13:05 crc kubenswrapper[4726]: I1001 16:13:05.125717 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eef09a6-cce6-45d8-ac7e-6b09f74d730a" containerName="extract" Oct 01 16:13:05 crc kubenswrapper[4726]: E1001 16:13:05.125729 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eef09a6-cce6-45d8-ac7e-6b09f74d730a" containerName="util" Oct 01 16:13:05 crc kubenswrapper[4726]: I1001 16:13:05.125737 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eef09a6-cce6-45d8-ac7e-6b09f74d730a" containerName="util" Oct 01 16:13:05 crc kubenswrapper[4726]: E1001 16:13:05.125759 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eef09a6-cce6-45d8-ac7e-6b09f74d730a" containerName="pull" Oct 01 16:13:05 crc kubenswrapper[4726]: I1001 16:13:05.125768 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eef09a6-cce6-45d8-ac7e-6b09f74d730a" containerName="pull" Oct 01 16:13:05 crc kubenswrapper[4726]: I1001 16:13:05.125912 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eef09a6-cce6-45d8-ac7e-6b09f74d730a" containerName="extract" Oct 01 16:13:05 crc kubenswrapper[4726]: I1001 16:13:05.126608 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6877df559-w8rr9" Oct 01 16:13:05 crc kubenswrapper[4726]: I1001 16:13:05.128237 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-lt4mp" Oct 01 16:13:05 crc kubenswrapper[4726]: I1001 16:13:05.133424 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99gt7\" (UniqueName: \"kubernetes.io/projected/07c756cd-913e-4237-a8eb-06dde60b6d4b-kube-api-access-99gt7\") pod \"openstack-operator-controller-operator-6877df559-w8rr9\" (UID: \"07c756cd-913e-4237-a8eb-06dde60b6d4b\") " pod="openstack-operators/openstack-operator-controller-operator-6877df559-w8rr9" Oct 01 16:13:05 crc kubenswrapper[4726]: I1001 16:13:05.155173 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6877df559-w8rr9"] Oct 01 16:13:05 crc kubenswrapper[4726]: I1001 16:13:05.234968 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99gt7\" (UniqueName: \"kubernetes.io/projected/07c756cd-913e-4237-a8eb-06dde60b6d4b-kube-api-access-99gt7\") pod \"openstack-operator-controller-operator-6877df559-w8rr9\" (UID: \"07c756cd-913e-4237-a8eb-06dde60b6d4b\") " pod="openstack-operators/openstack-operator-controller-operator-6877df559-w8rr9" Oct 01 16:13:05 crc kubenswrapper[4726]: I1001 16:13:05.257041 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99gt7\" (UniqueName: \"kubernetes.io/projected/07c756cd-913e-4237-a8eb-06dde60b6d4b-kube-api-access-99gt7\") pod \"openstack-operator-controller-operator-6877df559-w8rr9\" (UID: \"07c756cd-913e-4237-a8eb-06dde60b6d4b\") " pod="openstack-operators/openstack-operator-controller-operator-6877df559-w8rr9" Oct 01 16:13:05 crc kubenswrapper[4726]: I1001 16:13:05.449097 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6877df559-w8rr9" Oct 01 16:13:05 crc kubenswrapper[4726]: I1001 16:13:05.871019 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6877df559-w8rr9"] Oct 01 16:13:05 crc kubenswrapper[4726]: W1001 16:13:05.891473 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c756cd_913e_4237_a8eb_06dde60b6d4b.slice/crio-fea6670f10496ddb35237762c6a704c3a9a376d6c7ac26e6edacd1060788b7fe WatchSource:0}: Error finding container fea6670f10496ddb35237762c6a704c3a9a376d6c7ac26e6edacd1060788b7fe: Status 404 returned error can't find the container with id fea6670f10496ddb35237762c6a704c3a9a376d6c7ac26e6edacd1060788b7fe Oct 01 16:13:06 crc kubenswrapper[4726]: I1001 16:13:06.895932 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6877df559-w8rr9" event={"ID":"07c756cd-913e-4237-a8eb-06dde60b6d4b","Type":"ContainerStarted","Data":"fea6670f10496ddb35237762c6a704c3a9a376d6c7ac26e6edacd1060788b7fe"} Oct 01 16:13:09 crc kubenswrapper[4726]: I1001 16:13:09.915163 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6877df559-w8rr9" event={"ID":"07c756cd-913e-4237-a8eb-06dde60b6d4b","Type":"ContainerStarted","Data":"ed541194b3e5b70a7583c7e924762b4ae2318263698b9b3e24001de956afb44f"} Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.360832 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q5bzx"] Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.363308 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.379530 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q5bzx"] Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.554367 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnrrv\" (UniqueName: \"kubernetes.io/projected/9f2ddc53-446b-42b0-966a-f6b2f4eb6159-kube-api-access-rnrrv\") pod \"certified-operators-q5bzx\" (UID: \"9f2ddc53-446b-42b0-966a-f6b2f4eb6159\") " pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.554602 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f2ddc53-446b-42b0-966a-f6b2f4eb6159-utilities\") pod \"certified-operators-q5bzx\" (UID: \"9f2ddc53-446b-42b0-966a-f6b2f4eb6159\") " pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.554666 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f2ddc53-446b-42b0-966a-f6b2f4eb6159-catalog-content\") pod \"certified-operators-q5bzx\" (UID: \"9f2ddc53-446b-42b0-966a-f6b2f4eb6159\") " pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.655876 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnrrv\" (UniqueName: \"kubernetes.io/projected/9f2ddc53-446b-42b0-966a-f6b2f4eb6159-kube-api-access-rnrrv\") pod \"certified-operators-q5bzx\" (UID: \"9f2ddc53-446b-42b0-966a-f6b2f4eb6159\") " pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.655975 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f2ddc53-446b-42b0-966a-f6b2f4eb6159-utilities\") pod \"certified-operators-q5bzx\" (UID: \"9f2ddc53-446b-42b0-966a-f6b2f4eb6159\") " pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.655999 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f2ddc53-446b-42b0-966a-f6b2f4eb6159-catalog-content\") pod \"certified-operators-q5bzx\" (UID: \"9f2ddc53-446b-42b0-966a-f6b2f4eb6159\") " pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.656485 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f2ddc53-446b-42b0-966a-f6b2f4eb6159-utilities\") pod \"certified-operators-q5bzx\" (UID: \"9f2ddc53-446b-42b0-966a-f6b2f4eb6159\") " pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.656589 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f2ddc53-446b-42b0-966a-f6b2f4eb6159-catalog-content\") pod \"certified-operators-q5bzx\" (UID: \"9f2ddc53-446b-42b0-966a-f6b2f4eb6159\") " pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.689397 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnrrv\" (UniqueName: \"kubernetes.io/projected/9f2ddc53-446b-42b0-966a-f6b2f4eb6159-kube-api-access-rnrrv\") pod \"certified-operators-q5bzx\" (UID: \"9f2ddc53-446b-42b0-966a-f6b2f4eb6159\") " pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.936623 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6877df559-w8rr9" event={"ID":"07c756cd-913e-4237-a8eb-06dde60b6d4b","Type":"ContainerStarted","Data":"844ba5d9aba315f4adede527ccfb3ce83163eb71bd46c607555fc9529b123f9e"} Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.936968 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6877df559-w8rr9" Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.973135 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6877df559-w8rr9" podStartSLOduration=2.007641843 podStartE2EDuration="7.973121649s" podCreationTimestamp="2025-10-01 16:13:05 +0000 UTC" firstStartedPulling="2025-10-01 16:13:05.893287174 +0000 UTC m=+838.794839751" lastFinishedPulling="2025-10-01 16:13:11.85876698 +0000 UTC m=+844.760319557" observedRunningTime="2025-10-01 16:13:12.969989387 +0000 UTC m=+845.871541964" watchObservedRunningTime="2025-10-01 16:13:12.973121649 +0000 UTC m=+845.874674226" Oct 01 16:13:12 crc kubenswrapper[4726]: I1001 16:13:12.979862 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:13 crc kubenswrapper[4726]: I1001 16:13:13.398994 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q5bzx"] Oct 01 16:13:13 crc kubenswrapper[4726]: W1001 16:13:13.408036 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f2ddc53_446b_42b0_966a_f6b2f4eb6159.slice/crio-c6a1a7c984b6b5bdb2088d834c47356432c6e1a10ffc6761327604f7c97d9a79 WatchSource:0}: Error finding container c6a1a7c984b6b5bdb2088d834c47356432c6e1a10ffc6761327604f7c97d9a79: Status 404 returned error can't find the container with id c6a1a7c984b6b5bdb2088d834c47356432c6e1a10ffc6761327604f7c97d9a79 Oct 01 16:13:13 crc kubenswrapper[4726]: I1001 16:13:13.946860 4726 generic.go:334] "Generic (PLEG): container finished" podID="9f2ddc53-446b-42b0-966a-f6b2f4eb6159" containerID="dda9175c63681378090e090ff6816c82585e278e36f08511300cd3cdc526391c" exitCode=0 Oct 01 16:13:13 crc kubenswrapper[4726]: I1001 16:13:13.946927 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q5bzx" event={"ID":"9f2ddc53-446b-42b0-966a-f6b2f4eb6159","Type":"ContainerDied","Data":"dda9175c63681378090e090ff6816c82585e278e36f08511300cd3cdc526391c"} Oct 01 16:13:13 crc kubenswrapper[4726]: I1001 16:13:13.946999 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q5bzx" event={"ID":"9f2ddc53-446b-42b0-966a-f6b2f4eb6159","Type":"ContainerStarted","Data":"c6a1a7c984b6b5bdb2088d834c47356432c6e1a10ffc6761327604f7c97d9a79"} Oct 01 16:13:15 crc kubenswrapper[4726]: I1001 16:13:15.452390 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6877df559-w8rr9" Oct 01 16:13:18 crc kubenswrapper[4726]: I1001 16:13:18.983536 4726 generic.go:334] "Generic (PLEG): container finished" podID="9f2ddc53-446b-42b0-966a-f6b2f4eb6159" containerID="a9c570aa325f0be0eea13e66d78e5899f19016c0f113c0f478cb4c0d51ecb98d" exitCode=0 Oct 01 16:13:18 crc kubenswrapper[4726]: I1001 16:13:18.983653 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q5bzx" event={"ID":"9f2ddc53-446b-42b0-966a-f6b2f4eb6159","Type":"ContainerDied","Data":"a9c570aa325f0be0eea13e66d78e5899f19016c0f113c0f478cb4c0d51ecb98d"} Oct 01 16:13:19 crc kubenswrapper[4726]: I1001 16:13:19.995551 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q5bzx" event={"ID":"9f2ddc53-446b-42b0-966a-f6b2f4eb6159","Type":"ContainerStarted","Data":"b071c199680eda974c437330b04bc289212d3f95227441bd16f1d7341a47d832"} Oct 01 16:13:20 crc kubenswrapper[4726]: I1001 16:13:20.021473 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q5bzx" podStartSLOduration=2.313848322 podStartE2EDuration="8.021421488s" podCreationTimestamp="2025-10-01 16:13:12 +0000 UTC" firstStartedPulling="2025-10-01 16:13:13.952335567 +0000 UTC m=+846.853888184" lastFinishedPulling="2025-10-01 16:13:19.659908763 +0000 UTC m=+852.561461350" observedRunningTime="2025-10-01 16:13:20.016220775 +0000 UTC m=+852.917773412" watchObservedRunningTime="2025-10-01 16:13:20.021421488 +0000 UTC m=+852.922974105" Oct 01 16:13:22 crc kubenswrapper[4726]: I1001 16:13:22.980325 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:22 crc kubenswrapper[4726]: I1001 16:13:22.980572 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:23 crc kubenswrapper[4726]: I1001 16:13:23.020189 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.885902 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-zxd4h"] Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.887589 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-zxd4h" Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.889471 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-46gs7" Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.899005 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-wckhv"] Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.905314 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wckhv" Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.907741 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-sxdkc" Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.940164 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-zxd4h"] Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.950506 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdzhm\" (UniqueName: \"kubernetes.io/projected/117d55f2-eee8-454b-9091-3ffffd61c547-kube-api-access-cdzhm\") pod \"cinder-operator-controller-manager-644bddb6d8-zxd4h\" (UID: \"117d55f2-eee8-454b-9091-3ffffd61c547\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-zxd4h" Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.950567 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq7jm\" (UniqueName: \"kubernetes.io/projected/85f77f5d-6b74-40b5-b238-d3515b4a685c-kube-api-access-gq7jm\") pod \"barbican-operator-controller-manager-6ff8b75857-wckhv\" (UID: \"85f77f5d-6b74-40b5-b238-d3515b4a685c\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wckhv" Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.955126 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-wckhv"] Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.965111 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-wgkz2"] Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.966372 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-wgkz2" Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.976274 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-67rj9"] Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.977453 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-67rj9" Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.977539 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-56dk8" Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.988333 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-8qvwl" Oct 01 16:13:32 crc kubenswrapper[4726]: I1001 16:13:32.991088 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-wgkz2"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.019431 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-67rj9"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.038300 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-qlb92"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.039317 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-qlb92" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.046558 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-vl295" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.051568 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq7jm\" (UniqueName: \"kubernetes.io/projected/85f77f5d-6b74-40b5-b238-d3515b4a685c-kube-api-access-gq7jm\") pod \"barbican-operator-controller-manager-6ff8b75857-wckhv\" (UID: \"85f77f5d-6b74-40b5-b238-d3515b4a685c\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wckhv" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.051632 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d95v4\" (UniqueName: \"kubernetes.io/projected/daf7d005-ec55-4cdb-91d6-eac465e42e17-kube-api-access-d95v4\") pod \"glance-operator-controller-manager-84958c4d49-67rj9\" (UID: \"daf7d005-ec55-4cdb-91d6-eac465e42e17\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-67rj9" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.051675 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdzhm\" (UniqueName: \"kubernetes.io/projected/117d55f2-eee8-454b-9091-3ffffd61c547-kube-api-access-cdzhm\") pod \"cinder-operator-controller-manager-644bddb6d8-zxd4h\" (UID: \"117d55f2-eee8-454b-9091-3ffffd61c547\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-zxd4h" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.051700 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szdkt\" (UniqueName: \"kubernetes.io/projected/71d9bfcf-89ac-440f-b2b1-1c2969859b0d-kube-api-access-szdkt\") pod \"designate-operator-controller-manager-84f4f7b77b-wgkz2\" (UID: \"71d9bfcf-89ac-440f-b2b1-1c2969859b0d\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-wgkz2" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.058597 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-qlb92"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.070468 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q5bzx" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.071238 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-zjwvl"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.072202 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zjwvl" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.074631 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-tm2fh" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.081584 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdzhm\" (UniqueName: \"kubernetes.io/projected/117d55f2-eee8-454b-9091-3ffffd61c547-kube-api-access-cdzhm\") pod \"cinder-operator-controller-manager-644bddb6d8-zxd4h\" (UID: \"117d55f2-eee8-454b-9091-3ffffd61c547\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-zxd4h" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.083582 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq7jm\" (UniqueName: \"kubernetes.io/projected/85f77f5d-6b74-40b5-b238-d3515b4a685c-kube-api-access-gq7jm\") pod \"barbican-operator-controller-manager-6ff8b75857-wckhv\" (UID: \"85f77f5d-6b74-40b5-b238-d3515b4a685c\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wckhv" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.086120 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-zjwvl"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.113090 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.114232 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.121442 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-rk7rf" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.121619 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.122113 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-cpbbq"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.123113 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-cpbbq" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.129632 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-6qm68" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.137124 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-qd8pc"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.138148 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-qd8pc" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.144932 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.144968 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-5hvl7"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.145771 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-5hvl7" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.146520 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-jvxps" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.147567 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.150922 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-gsbcl" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.152417 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d95v4\" (UniqueName: \"kubernetes.io/projected/daf7d005-ec55-4cdb-91d6-eac465e42e17-kube-api-access-d95v4\") pod \"glance-operator-controller-manager-84958c4d49-67rj9\" (UID: \"daf7d005-ec55-4cdb-91d6-eac465e42e17\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-67rj9" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.152486 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k696c\" (UniqueName: \"kubernetes.io/projected/f82a5172-6387-4a24-a7e4-a5bd4c90d310-kube-api-access-k696c\") pod \"heat-operator-controller-manager-5d889d78cf-qlb92\" (UID: \"f82a5172-6387-4a24-a7e4-a5bd4c90d310\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-qlb92" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.152528 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv8jq\" (UniqueName: \"kubernetes.io/projected/39860f7b-9a7f-44b1-8bf4-7a2204092c18-kube-api-access-lv8jq\") pod \"horizon-operator-controller-manager-9f4696d94-zjwvl\" (UID: \"39860f7b-9a7f-44b1-8bf4-7a2204092c18\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zjwvl" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.152552 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szdkt\" (UniqueName: \"kubernetes.io/projected/71d9bfcf-89ac-440f-b2b1-1c2969859b0d-kube-api-access-szdkt\") pod \"designate-operator-controller-manager-84f4f7b77b-wgkz2\" (UID: \"71d9bfcf-89ac-440f-b2b1-1c2969859b0d\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-wgkz2" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.158612 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.166442 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-q4hj4" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.176899 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d95v4\" (UniqueName: \"kubernetes.io/projected/daf7d005-ec55-4cdb-91d6-eac465e42e17-kube-api-access-d95v4\") pod \"glance-operator-controller-manager-84958c4d49-67rj9\" (UID: \"daf7d005-ec55-4cdb-91d6-eac465e42e17\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-67rj9" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.178942 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.193031 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szdkt\" (UniqueName: \"kubernetes.io/projected/71d9bfcf-89ac-440f-b2b1-1c2969859b0d-kube-api-access-szdkt\") pod \"designate-operator-controller-manager-84f4f7b77b-wgkz2\" (UID: \"71d9bfcf-89ac-440f-b2b1-1c2969859b0d\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-wgkz2" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.198587 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-cpbbq"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.203280 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-5hvl7"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.221188 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-zxd4h" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.239853 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wckhv" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.246855 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.247917 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.250147 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-t4jtq" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.257961 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzjkz\" (UniqueName: \"kubernetes.io/projected/6b63c738-b108-4193-9c9d-11f4eb8227aa-kube-api-access-nzjkz\") pod \"infra-operator-controller-manager-9d6c5db85-z5lfd\" (UID: \"6b63c738-b108-4193-9c9d-11f4eb8227aa\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.258007 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8xvl\" (UniqueName: \"kubernetes.io/projected/8e085c54-533c-4186-a6fe-5d8a5ccbac2f-kube-api-access-r8xvl\") pod \"ironic-operator-controller-manager-5cd4858477-cpbbq\" (UID: \"8e085c54-533c-4186-a6fe-5d8a5ccbac2f\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-cpbbq" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.258032 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cvcd\" (UniqueName: \"kubernetes.io/projected/9f6777bb-5041-4c66-967a-069fa217bf99-kube-api-access-6cvcd\") pod \"mariadb-operator-controller-manager-88c7-5vlmb\" (UID: \"9f6777bb-5041-4c66-967a-069fa217bf99\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.258081 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k696c\" (UniqueName: \"kubernetes.io/projected/f82a5172-6387-4a24-a7e4-a5bd4c90d310-kube-api-access-k696c\") pod \"heat-operator-controller-manager-5d889d78cf-qlb92\" (UID: \"f82a5172-6387-4a24-a7e4-a5bd4c90d310\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-qlb92" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.258119 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv8jq\" (UniqueName: \"kubernetes.io/projected/39860f7b-9a7f-44b1-8bf4-7a2204092c18-kube-api-access-lv8jq\") pod \"horizon-operator-controller-manager-9f4696d94-zjwvl\" (UID: \"39860f7b-9a7f-44b1-8bf4-7a2204092c18\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zjwvl" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.258146 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvgpp\" (UniqueName: \"kubernetes.io/projected/23922646-bd87-424b-85d7-8a2a55056438-kube-api-access-xvgpp\") pod \"keystone-operator-controller-manager-5bd55b4bff-5hvl7\" (UID: \"23922646-bd87-424b-85d7-8a2a55056438\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-5hvl7" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.258169 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6b63c738-b108-4193-9c9d-11f4eb8227aa-cert\") pod \"infra-operator-controller-manager-9d6c5db85-z5lfd\" (UID: \"6b63c738-b108-4193-9c9d-11f4eb8227aa\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.258186 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bscl\" (UniqueName: \"kubernetes.io/projected/2d90550d-aedc-4e50-81f9-da6f285f8c2b-kube-api-access-2bscl\") pod \"manila-operator-controller-manager-6d68dbc695-qd8pc\" (UID: \"2d90550d-aedc-4e50-81f9-da6f285f8c2b\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-qd8pc" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.258668 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.263386 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-qd8pc"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.266545 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.273389 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-2kkkp" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.276737 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-t2vwz"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.277745 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-t2vwz" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.279027 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-k8qvr" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.283795 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv8jq\" (UniqueName: \"kubernetes.io/projected/39860f7b-9a7f-44b1-8bf4-7a2204092c18-kube-api-access-lv8jq\") pod \"horizon-operator-controller-manager-9f4696d94-zjwvl\" (UID: \"39860f7b-9a7f-44b1-8bf4-7a2204092c18\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zjwvl" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.285402 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-wgkz2" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.287217 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k696c\" (UniqueName: \"kubernetes.io/projected/f82a5172-6387-4a24-a7e4-a5bd4c90d310-kube-api-access-k696c\") pod \"heat-operator-controller-manager-5d889d78cf-qlb92\" (UID: \"f82a5172-6387-4a24-a7e4-a5bd4c90d310\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-qlb92" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.303102 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.309507 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-67rj9" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.311368 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.315725 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-t2vwz"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.328542 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-7s6mh"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.341701 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-7s6mh" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.346016 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-lgmqj" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.360909 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6x4x\" (UniqueName: \"kubernetes.io/projected/0d89451f-e13a-4195-a2b2-f16443c43463-kube-api-access-r6x4x\") pod \"neutron-operator-controller-manager-849d5b9b84-w9xkn\" (UID: \"0d89451f-e13a-4195-a2b2-f16443c43463\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.360961 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bscl\" (UniqueName: \"kubernetes.io/projected/2d90550d-aedc-4e50-81f9-da6f285f8c2b-kube-api-access-2bscl\") pod \"manila-operator-controller-manager-6d68dbc695-qd8pc\" (UID: \"2d90550d-aedc-4e50-81f9-da6f285f8c2b\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-qd8pc" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.361006 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzjkz\" (UniqueName: \"kubernetes.io/projected/6b63c738-b108-4193-9c9d-11f4eb8227aa-kube-api-access-nzjkz\") pod \"infra-operator-controller-manager-9d6c5db85-z5lfd\" (UID: \"6b63c738-b108-4193-9c9d-11f4eb8227aa\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.361069 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8xvl\" (UniqueName: \"kubernetes.io/projected/8e085c54-533c-4186-a6fe-5d8a5ccbac2f-kube-api-access-r8xvl\") pod \"ironic-operator-controller-manager-5cd4858477-cpbbq\" (UID: \"8e085c54-533c-4186-a6fe-5d8a5ccbac2f\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-cpbbq" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.361110 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cvcd\" (UniqueName: \"kubernetes.io/projected/9f6777bb-5041-4c66-967a-069fa217bf99-kube-api-access-6cvcd\") pod \"mariadb-operator-controller-manager-88c7-5vlmb\" (UID: \"9f6777bb-5041-4c66-967a-069fa217bf99\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.361153 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98dl7\" (UniqueName: \"kubernetes.io/projected/b9f1f46c-c6fa-443f-b53d-54614779c0cc-kube-api-access-98dl7\") pod \"octavia-operator-controller-manager-7b787867f4-t2vwz\" (UID: \"b9f1f46c-c6fa-443f-b53d-54614779c0cc\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-t2vwz" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.361228 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvgpp\" (UniqueName: \"kubernetes.io/projected/23922646-bd87-424b-85d7-8a2a55056438-kube-api-access-xvgpp\") pod \"keystone-operator-controller-manager-5bd55b4bff-5hvl7\" (UID: \"23922646-bd87-424b-85d7-8a2a55056438\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-5hvl7" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.361262 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh7gt\" (UniqueName: \"kubernetes.io/projected/0c8fbaaf-4ce0-4388-855e-fb85534f05e6-kube-api-access-wh7gt\") pod \"nova-operator-controller-manager-64cd67b5cb-zdkvq\" (UID: \"0c8fbaaf-4ce0-4388-855e-fb85534f05e6\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.361297 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6b63c738-b108-4193-9c9d-11f4eb8227aa-cert\") pod \"infra-operator-controller-manager-9d6c5db85-z5lfd\" (UID: \"6b63c738-b108-4193-9c9d-11f4eb8227aa\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" Oct 01 16:13:33 crc kubenswrapper[4726]: E1001 16:13:33.361460 4726 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 01 16:13:33 crc kubenswrapper[4726]: E1001 16:13:33.361528 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b63c738-b108-4193-9c9d-11f4eb8227aa-cert podName:6b63c738-b108-4193-9c9d-11f4eb8227aa nodeName:}" failed. No retries permitted until 2025-10-01 16:13:33.861498061 +0000 UTC m=+866.763050638 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6b63c738-b108-4193-9c9d-11f4eb8227aa-cert") pod "infra-operator-controller-manager-9d6c5db85-z5lfd" (UID: "6b63c738-b108-4193-9c9d-11f4eb8227aa") : secret "infra-operator-webhook-server-cert" not found Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.366829 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-qlb92" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.389439 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cvcd\" (UniqueName: \"kubernetes.io/projected/9f6777bb-5041-4c66-967a-069fa217bf99-kube-api-access-6cvcd\") pod \"mariadb-operator-controller-manager-88c7-5vlmb\" (UID: \"9f6777bb-5041-4c66-967a-069fa217bf99\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.395114 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8xvl\" (UniqueName: \"kubernetes.io/projected/8e085c54-533c-4186-a6fe-5d8a5ccbac2f-kube-api-access-r8xvl\") pod \"ironic-operator-controller-manager-5cd4858477-cpbbq\" (UID: \"8e085c54-533c-4186-a6fe-5d8a5ccbac2f\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-cpbbq" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.395188 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.396719 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.399562 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bscl\" (UniqueName: \"kubernetes.io/projected/2d90550d-aedc-4e50-81f9-da6f285f8c2b-kube-api-access-2bscl\") pod \"manila-operator-controller-manager-6d68dbc695-qd8pc\" (UID: \"2d90550d-aedc-4e50-81f9-da6f285f8c2b\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-qd8pc" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.399887 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-d4xpd" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.403607 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvgpp\" (UniqueName: \"kubernetes.io/projected/23922646-bd87-424b-85d7-8a2a55056438-kube-api-access-xvgpp\") pod \"keystone-operator-controller-manager-5bd55b4bff-5hvl7\" (UID: \"23922646-bd87-424b-85d7-8a2a55056438\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-5hvl7" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.407649 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzjkz\" (UniqueName: \"kubernetes.io/projected/6b63c738-b108-4193-9c9d-11f4eb8227aa-kube-api-access-nzjkz\") pod \"infra-operator-controller-manager-9d6c5db85-z5lfd\" (UID: \"6b63c738-b108-4193-9c9d-11f4eb8227aa\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.407712 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.408766 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.410917 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.411773 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-lj24x" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.425427 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-pl7jv"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.427547 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-pl7jv" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.433676 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-7s6mh"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.435364 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zjwvl" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.436031 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-sq2dz" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.440427 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.449599 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-pl7jv"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.459158 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.468577 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.469923 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.472981 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.473648 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh7gt\" (UniqueName: \"kubernetes.io/projected/0c8fbaaf-4ce0-4388-855e-fb85534f05e6-kube-api-access-wh7gt\") pod \"nova-operator-controller-manager-64cd67b5cb-zdkvq\" (UID: \"0c8fbaaf-4ce0-4388-855e-fb85534f05e6\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.473688 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/420c55b8-4442-41fa-98a4-338f7338ceb4-cert\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf\" (UID: \"420c55b8-4442-41fa-98a4-338f7338ceb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.473736 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6x4x\" (UniqueName: \"kubernetes.io/projected/0d89451f-e13a-4195-a2b2-f16443c43463-kube-api-access-r6x4x\") pod \"neutron-operator-controller-manager-849d5b9b84-w9xkn\" (UID: \"0d89451f-e13a-4195-a2b2-f16443c43463\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.473787 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnnsm\" (UniqueName: \"kubernetes.io/projected/420c55b8-4442-41fa-98a4-338f7338ceb4-kube-api-access-bnnsm\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf\" (UID: \"420c55b8-4442-41fa-98a4-338f7338ceb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.473827 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6rcc\" (UniqueName: \"kubernetes.io/projected/1845e7cd-8784-4994-8f4f-3015bf7d6b3b-kube-api-access-d6rcc\") pod \"ovn-operator-controller-manager-9976ff44c-j4t5g\" (UID: \"1845e7cd-8784-4994-8f4f-3015bf7d6b3b\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.473853 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98dl7\" (UniqueName: \"kubernetes.io/projected/b9f1f46c-c6fa-443f-b53d-54614779c0cc-kube-api-access-98dl7\") pod \"octavia-operator-controller-manager-7b787867f4-t2vwz\" (UID: \"b9f1f46c-c6fa-443f-b53d-54614779c0cc\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-t2vwz" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.473875 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hjwg\" (UniqueName: \"kubernetes.io/projected/ff753ea2-fa5b-4367-b476-596fd49d0557-kube-api-access-7hjwg\") pod \"placement-operator-controller-manager-589c58c6c-7s6mh\" (UID: \"ff753ea2-fa5b-4367-b476-596fd49d0557\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-7s6mh" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.474303 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-nzxsh" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.479265 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-cpbbq" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.504222 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6x4x\" (UniqueName: \"kubernetes.io/projected/0d89451f-e13a-4195-a2b2-f16443c43463-kube-api-access-r6x4x\") pod \"neutron-operator-controller-manager-849d5b9b84-w9xkn\" (UID: \"0d89451f-e13a-4195-a2b2-f16443c43463\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.507111 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98dl7\" (UniqueName: \"kubernetes.io/projected/b9f1f46c-c6fa-443f-b53d-54614779c0cc-kube-api-access-98dl7\") pod \"octavia-operator-controller-manager-7b787867f4-t2vwz\" (UID: \"b9f1f46c-c6fa-443f-b53d-54614779c0cc\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-t2vwz" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.513099 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh7gt\" (UniqueName: \"kubernetes.io/projected/0c8fbaaf-4ce0-4388-855e-fb85534f05e6-kube-api-access-wh7gt\") pod \"nova-operator-controller-manager-64cd67b5cb-zdkvq\" (UID: \"0c8fbaaf-4ce0-4388-855e-fb85534f05e6\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.530587 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-nv4ww"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.531791 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-nv4ww" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.534713 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-tr9bc" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.552930 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-nv4ww"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.553819 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-qd8pc" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.575708 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/420c55b8-4442-41fa-98a4-338f7338ceb4-cert\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf\" (UID: \"420c55b8-4442-41fa-98a4-338f7338ceb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.575769 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q4dn\" (UniqueName: \"kubernetes.io/projected/fa99155f-85e2-406e-84c4-4a95ab74407f-kube-api-access-2q4dn\") pod \"swift-operator-controller-manager-84d6b4b759-pl7jv\" (UID: \"fa99155f-85e2-406e-84c4-4a95ab74407f\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-pl7jv" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.575806 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnnsm\" (UniqueName: \"kubernetes.io/projected/420c55b8-4442-41fa-98a4-338f7338ceb4-kube-api-access-bnnsm\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf\" (UID: \"420c55b8-4442-41fa-98a4-338f7338ceb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.575846 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6rcc\" (UniqueName: \"kubernetes.io/projected/1845e7cd-8784-4994-8f4f-3015bf7d6b3b-kube-api-access-d6rcc\") pod \"ovn-operator-controller-manager-9976ff44c-j4t5g\" (UID: \"1845e7cd-8784-4994-8f4f-3015bf7d6b3b\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.575868 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hjwg\" (UniqueName: \"kubernetes.io/projected/ff753ea2-fa5b-4367-b476-596fd49d0557-kube-api-access-7hjwg\") pod \"placement-operator-controller-manager-589c58c6c-7s6mh\" (UID: \"ff753ea2-fa5b-4367-b476-596fd49d0557\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-7s6mh" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.575905 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9hbb\" (UniqueName: \"kubernetes.io/projected/75c9f506-beb2-4fe6-9ada-3ed684131102-kube-api-access-p9hbb\") pod \"telemetry-operator-controller-manager-7866c8b47c-z6rlg\" (UID: \"75c9f506-beb2-4fe6-9ada-3ed684131102\") " pod="openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg" Oct 01 16:13:33 crc kubenswrapper[4726]: E1001 16:13:33.576086 4726 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 01 16:13:33 crc kubenswrapper[4726]: E1001 16:13:33.576129 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/420c55b8-4442-41fa-98a4-338f7338ceb4-cert podName:420c55b8-4442-41fa-98a4-338f7338ceb4 nodeName:}" failed. No retries permitted until 2025-10-01 16:13:34.076114159 +0000 UTC m=+866.977666736 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/420c55b8-4442-41fa-98a4-338f7338ceb4-cert") pod "openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" (UID: "420c55b8-4442-41fa-98a4-338f7338ceb4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.583344 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-9nqm2"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.584556 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9nqm2" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.588294 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-5hvl7" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.594491 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-mpm7z" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.594658 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-9nqm2"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.610181 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.627715 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.629431 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnnsm\" (UniqueName: \"kubernetes.io/projected/420c55b8-4442-41fa-98a4-338f7338ceb4-kube-api-access-bnnsm\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf\" (UID: \"420c55b8-4442-41fa-98a4-338f7338ceb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.637510 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.637707 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hjwg\" (UniqueName: \"kubernetes.io/projected/ff753ea2-fa5b-4367-b476-596fd49d0557-kube-api-access-7hjwg\") pod \"placement-operator-controller-manager-589c58c6c-7s6mh\" (UID: \"ff753ea2-fa5b-4367-b476-596fd49d0557\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-7s6mh" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.638021 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6rcc\" (UniqueName: \"kubernetes.io/projected/1845e7cd-8784-4994-8f4f-3015bf7d6b3b-kube-api-access-d6rcc\") pod \"ovn-operator-controller-manager-9976ff44c-j4t5g\" (UID: \"1845e7cd-8784-4994-8f4f-3015bf7d6b3b\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.647778 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-t2vwz" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.656854 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.657904 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.666572 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.666822 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-b5rk7" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.666854 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-7s6mh" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.678690 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9hbb\" (UniqueName: \"kubernetes.io/projected/75c9f506-beb2-4fe6-9ada-3ed684131102-kube-api-access-p9hbb\") pod \"telemetry-operator-controller-manager-7866c8b47c-z6rlg\" (UID: \"75c9f506-beb2-4fe6-9ada-3ed684131102\") " pod="openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.678741 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lwb5\" (UniqueName: \"kubernetes.io/projected/b824be4d-8e37-42b5-8fe2-0dc6a1c4b4a1-kube-api-access-4lwb5\") pod \"watcher-operator-controller-manager-6b9957f54f-9nqm2\" (UID: \"b824be4d-8e37-42b5-8fe2-0dc6a1c4b4a1\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9nqm2" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.678786 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx9nh\" (UniqueName: \"kubernetes.io/projected/0396f184-2150-4944-a81c-c0d0fafe1317-kube-api-access-tx9nh\") pod \"test-operator-controller-manager-85777745bb-nv4ww\" (UID: \"0396f184-2150-4944-a81c-c0d0fafe1317\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-nv4ww" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.678817 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q4dn\" (UniqueName: \"kubernetes.io/projected/fa99155f-85e2-406e-84c4-4a95ab74407f-kube-api-access-2q4dn\") pod \"swift-operator-controller-manager-84d6b4b759-pl7jv\" (UID: \"fa99155f-85e2-406e-84c4-4a95ab74407f\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-pl7jv" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.698402 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.725958 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.753731 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9hbb\" (UniqueName: \"kubernetes.io/projected/75c9f506-beb2-4fe6-9ada-3ed684131102-kube-api-access-p9hbb\") pod \"telemetry-operator-controller-manager-7866c8b47c-z6rlg\" (UID: \"75c9f506-beb2-4fe6-9ada-3ed684131102\") " pod="openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.754310 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q4dn\" (UniqueName: \"kubernetes.io/projected/fa99155f-85e2-406e-84c4-4a95ab74407f-kube-api-access-2q4dn\") pod \"swift-operator-controller-manager-84d6b4b759-pl7jv\" (UID: \"fa99155f-85e2-406e-84c4-4a95ab74407f\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-pl7jv" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.771042 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.771969 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.774687 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-pl7jv" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.775173 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-mjcr8" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.777368 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv"] Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.781555 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lwb5\" (UniqueName: \"kubernetes.io/projected/b824be4d-8e37-42b5-8fe2-0dc6a1c4b4a1-kube-api-access-4lwb5\") pod \"watcher-operator-controller-manager-6b9957f54f-9nqm2\" (UID: \"b824be4d-8e37-42b5-8fe2-0dc6a1c4b4a1\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9nqm2" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.781610 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42f6c148-bcc5-4054-88da-dc79d49baeff-cert\") pod \"openstack-operator-controller-manager-5f569c4895-tl4ld\" (UID: \"42f6c148-bcc5-4054-88da-dc79d49baeff\") " pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.781665 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx9nh\" (UniqueName: \"kubernetes.io/projected/0396f184-2150-4944-a81c-c0d0fafe1317-kube-api-access-tx9nh\") pod \"test-operator-controller-manager-85777745bb-nv4ww\" (UID: \"0396f184-2150-4944-a81c-c0d0fafe1317\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-nv4ww" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.781735 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cw5h\" (UniqueName: \"kubernetes.io/projected/42f6c148-bcc5-4054-88da-dc79d49baeff-kube-api-access-9cw5h\") pod \"openstack-operator-controller-manager-5f569c4895-tl4ld\" (UID: \"42f6c148-bcc5-4054-88da-dc79d49baeff\") " pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.812605 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lwb5\" (UniqueName: \"kubernetes.io/projected/b824be4d-8e37-42b5-8fe2-0dc6a1c4b4a1-kube-api-access-4lwb5\") pod \"watcher-operator-controller-manager-6b9957f54f-9nqm2\" (UID: \"b824be4d-8e37-42b5-8fe2-0dc6a1c4b4a1\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9nqm2" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.812910 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.828609 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx9nh\" (UniqueName: \"kubernetes.io/projected/0396f184-2150-4944-a81c-c0d0fafe1317-kube-api-access-tx9nh\") pod \"test-operator-controller-manager-85777745bb-nv4ww\" (UID: \"0396f184-2150-4944-a81c-c0d0fafe1317\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-nv4ww" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.878580 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9nqm2" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.878951 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-nv4ww" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.888631 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6b63c738-b108-4193-9c9d-11f4eb8227aa-cert\") pod \"infra-operator-controller-manager-9d6c5db85-z5lfd\" (UID: \"6b63c738-b108-4193-9c9d-11f4eb8227aa\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.888710 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cw5h\" (UniqueName: \"kubernetes.io/projected/42f6c148-bcc5-4054-88da-dc79d49baeff-kube-api-access-9cw5h\") pod \"openstack-operator-controller-manager-5f569c4895-tl4ld\" (UID: \"42f6c148-bcc5-4054-88da-dc79d49baeff\") " pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.888763 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnj8l\" (UniqueName: \"kubernetes.io/projected/965ea363-f11c-4ced-9988-debe220ee3f9-kube-api-access-tnj8l\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv\" (UID: \"965ea363-f11c-4ced-9988-debe220ee3f9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.888800 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42f6c148-bcc5-4054-88da-dc79d49baeff-cert\") pod \"openstack-operator-controller-manager-5f569c4895-tl4ld\" (UID: \"42f6c148-bcc5-4054-88da-dc79d49baeff\") " pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" Oct 01 16:13:33 crc kubenswrapper[4726]: E1001 16:13:33.888962 4726 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 01 16:13:33 crc kubenswrapper[4726]: E1001 16:13:33.889015 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/42f6c148-bcc5-4054-88da-dc79d49baeff-cert podName:42f6c148-bcc5-4054-88da-dc79d49baeff nodeName:}" failed. No retries permitted until 2025-10-01 16:13:34.388999334 +0000 UTC m=+867.290551911 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/42f6c148-bcc5-4054-88da-dc79d49baeff-cert") pod "openstack-operator-controller-manager-5f569c4895-tl4ld" (UID: "42f6c148-bcc5-4054-88da-dc79d49baeff") : secret "webhook-server-cert" not found Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.903376 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6b63c738-b108-4193-9c9d-11f4eb8227aa-cert\") pod \"infra-operator-controller-manager-9d6c5db85-z5lfd\" (UID: \"6b63c738-b108-4193-9c9d-11f4eb8227aa\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.936535 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cw5h\" (UniqueName: \"kubernetes.io/projected/42f6c148-bcc5-4054-88da-dc79d49baeff-kube-api-access-9cw5h\") pod \"openstack-operator-controller-manager-5f569c4895-tl4ld\" (UID: \"42f6c148-bcc5-4054-88da-dc79d49baeff\") " pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" Oct 01 16:13:33 crc kubenswrapper[4726]: I1001 16:13:33.990084 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnj8l\" (UniqueName: \"kubernetes.io/projected/965ea363-f11c-4ced-9988-debe220ee3f9-kube-api-access-tnj8l\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv\" (UID: \"965ea363-f11c-4ced-9988-debe220ee3f9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv" Oct 01 16:13:34 crc kubenswrapper[4726]: I1001 16:13:34.017478 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnj8l\" (UniqueName: \"kubernetes.io/projected/965ea363-f11c-4ced-9988-debe220ee3f9-kube-api-access-tnj8l\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv\" (UID: \"965ea363-f11c-4ced-9988-debe220ee3f9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv" Oct 01 16:13:34 crc kubenswrapper[4726]: I1001 16:13:34.068462 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" Oct 01 16:13:34 crc kubenswrapper[4726]: I1001 16:13:34.088191 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-zxd4h"] Oct 01 16:13:34 crc kubenswrapper[4726]: I1001 16:13:34.110650 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/420c55b8-4442-41fa-98a4-338f7338ceb4-cert\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf\" (UID: \"420c55b8-4442-41fa-98a4-338f7338ceb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" Oct 01 16:13:34 crc kubenswrapper[4726]: I1001 16:13:34.120814 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/420c55b8-4442-41fa-98a4-338f7338ceb4-cert\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf\" (UID: \"420c55b8-4442-41fa-98a4-338f7338ceb4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" Oct 01 16:13:34 crc kubenswrapper[4726]: I1001 16:13:34.234837 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv" Oct 01 16:13:34 crc kubenswrapper[4726]: I1001 16:13:34.340244 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" Oct 01 16:13:34 crc kubenswrapper[4726]: I1001 16:13:34.413707 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42f6c148-bcc5-4054-88da-dc79d49baeff-cert\") pod \"openstack-operator-controller-manager-5f569c4895-tl4ld\" (UID: \"42f6c148-bcc5-4054-88da-dc79d49baeff\") " pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" Oct 01 16:13:34 crc kubenswrapper[4726]: E1001 16:13:34.413891 4726 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 01 16:13:34 crc kubenswrapper[4726]: E1001 16:13:34.413945 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/42f6c148-bcc5-4054-88da-dc79d49baeff-cert podName:42f6c148-bcc5-4054-88da-dc79d49baeff nodeName:}" failed. No retries permitted until 2025-10-01 16:13:35.413931531 +0000 UTC m=+868.315484098 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/42f6c148-bcc5-4054-88da-dc79d49baeff-cert") pod "openstack-operator-controller-manager-5f569c4895-tl4ld" (UID: "42f6c148-bcc5-4054-88da-dc79d49baeff") : secret "webhook-server-cert" not found Oct 01 16:13:34 crc kubenswrapper[4726]: I1001 16:13:34.426808 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-wgkz2"] Oct 01 16:13:34 crc kubenswrapper[4726]: I1001 16:13:34.440153 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-67rj9"] Oct 01 16:13:34 crc kubenswrapper[4726]: I1001 16:13:34.471266 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-wckhv"] Oct 01 16:13:34 crc kubenswrapper[4726]: I1001 16:13:34.631134 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-zjwvl"] Oct 01 16:13:34 crc kubenswrapper[4726]: I1001 16:13:34.708508 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-qlb92"] Oct 01 16:13:34 crc kubenswrapper[4726]: I1001 16:13:34.994414 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-qd8pc"] Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.000671 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-nv4ww"] Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.006090 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q5bzx"] Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.016395 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-9nqm2"] Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.036808 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-5hvl7"] Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.049325 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg"] Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.058076 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-pl7jv"] Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.063203 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-cpbbq"] Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.067223 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-t2vwz"] Oct 01 16:13:35 crc kubenswrapper[4726]: W1001 16:13:35.071571 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9f1f46c_c6fa_443f_b53d_54614779c0cc.slice/crio-230e84ad9ead8181001b73b6798d1636dceebe60cabaf1cf08f9b5aff987452d WatchSource:0}: Error finding container 230e84ad9ead8181001b73b6798d1636dceebe60cabaf1cf08f9b5aff987452d: Status 404 returned error can't find the container with id 230e84ad9ead8181001b73b6798d1636dceebe60cabaf1cf08f9b5aff987452d Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.072444 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-7s6mh"] Oct 01 16:13:35 crc kubenswrapper[4726]: E1001 16:13:35.079065 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d6rcc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-9976ff44c-j4t5g_openstack-operators(1845e7cd-8784-4994-8f4f-3015bf7d6b3b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:13:35 crc kubenswrapper[4726]: E1001 16:13:35.079137 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.163:5001/openstack-k8s-operators/telemetry-operator:020bd70758b454f4f03f761350467eb9337e2b6e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p9hbb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-7866c8b47c-z6rlg_openstack-operators(75c9f506-beb2-4fe6-9ada-3ed684131102): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:13:35 crc kubenswrapper[4726]: E1001 16:13:35.083689 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wh7gt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-64cd67b5cb-zdkvq_openstack-operators(0c8fbaaf-4ce0-4388-855e-fb85534f05e6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:13:35 crc kubenswrapper[4726]: E1001 16:13:35.087502 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:15d7b5a365350a831ca59d984df67fadeccf89d599e487a7597b105afb82ce4a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6cvcd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-88c7-5vlmb_openstack-operators(9f6777bb-5041-4c66-967a-069fa217bf99): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.088592 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb"] Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.103424 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq"] Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.108651 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wckhv" event={"ID":"85f77f5d-6b74-40b5-b238-d3515b4a685c","Type":"ContainerStarted","Data":"7b595e7976d9e4b4338ba7c3f718d61250e0e840e9851ce315929cc68eb557c1"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.109124 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g"] Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.110044 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g" event={"ID":"1845e7cd-8784-4994-8f4f-3015bf7d6b3b","Type":"ContainerStarted","Data":"f64022e9287a558232e71eb948f17d30874424db7fc1bab3457cf50aca514603"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.115560 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-cpbbq" event={"ID":"8e085c54-533c-4186-a6fe-5d8a5ccbac2f","Type":"ContainerStarted","Data":"b152fb969295dd6b74a941ed91fb950bb3c070196981ca3c47a050baa265ecbf"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.116949 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-7s6mh" event={"ID":"ff753ea2-fa5b-4367-b476-596fd49d0557","Type":"ContainerStarted","Data":"5215bb16c7fe9074be4a5fb9456834e4fd389ac580cb7f10a281e794eeff6769"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.118192 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-t2vwz" event={"ID":"b9f1f46c-c6fa-443f-b53d-54614779c0cc","Type":"ContainerStarted","Data":"230e84ad9ead8181001b73b6798d1636dceebe60cabaf1cf08f9b5aff987452d"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.119298 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-5hvl7" event={"ID":"23922646-bd87-424b-85d7-8a2a55056438","Type":"ContainerStarted","Data":"fa51258a99abf8c45a096824e423f48e7e5737de0b3bff9c6d8b631aa9d757a0"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.120321 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9nqm2" event={"ID":"b824be4d-8e37-42b5-8fe2-0dc6a1c4b4a1","Type":"ContainerStarted","Data":"59b519e8e24fce1b8965631eae1d8b0c684ee5ae7ae2b9b34885f69018a63100"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.123439 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb" event={"ID":"9f6777bb-5041-4c66-967a-069fa217bf99","Type":"ContainerStarted","Data":"a5f6d6413c68c7d373299d6099c0e7ce0955d11c83f2e951699fd9e171999e0d"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.125194 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-67rj9" event={"ID":"daf7d005-ec55-4cdb-91d6-eac465e42e17","Type":"ContainerStarted","Data":"d1486e803ec54cbed890ecc890a946eca5a4ca4b1165fb491872b22597afc9ff"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.126276 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zjwvl" event={"ID":"39860f7b-9a7f-44b1-8bf4-7a2204092c18","Type":"ContainerStarted","Data":"bdfa6acec31fc6b445683cf5291bc834715d73cc7b2633a17f742bb429844c9a"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.127215 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-zxd4h" event={"ID":"117d55f2-eee8-454b-9091-3ffffd61c547","Type":"ContainerStarted","Data":"5a7a9fe7b737a4ce3e2ce4d908b39a7ec5842998173030169e9aa71ea93deae5"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.128284 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg" event={"ID":"75c9f506-beb2-4fe6-9ada-3ed684131102","Type":"ContainerStarted","Data":"33e5777e477c1bc94dcac3cc9c6d124f52b1405c6748a3e3626eff83cea81b0b"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.129210 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-wgkz2" event={"ID":"71d9bfcf-89ac-440f-b2b1-1c2969859b0d","Type":"ContainerStarted","Data":"9f9ab915dc8eaa89b2c9f6ab9666aa2491d397ec8f0baab961cb346760f85031"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.130657 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-qd8pc" event={"ID":"2d90550d-aedc-4e50-81f9-da6f285f8c2b","Type":"ContainerStarted","Data":"43f80c7958ab6338dbd50acc64cb9d421208312b0c4e41ab9187ed36053fdf0c"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.131307 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-pl7jv" event={"ID":"fa99155f-85e2-406e-84c4-4a95ab74407f","Type":"ContainerStarted","Data":"d9fbcbb2336b9d0b72474d0f8bd4361a921faec20b6757523d3f13b887d13d27"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.132139 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-nv4ww" event={"ID":"0396f184-2150-4944-a81c-c0d0fafe1317","Type":"ContainerStarted","Data":"f780b79866441c4c1a37918acbec1c660ca294c7f6500eaf7edfaf45ff1b03ed"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.132894 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-qlb92" event={"ID":"f82a5172-6387-4a24-a7e4-a5bd4c90d310","Type":"ContainerStarted","Data":"e54f465eaf34fa0a81eb19b96e8d511c62bab80863fecc4deb8919d4f28c70fe"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.134392 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq" event={"ID":"0c8fbaaf-4ce0-4388-855e-fb85534f05e6","Type":"ContainerStarted","Data":"6ef0fef9e5028ea7455a4bce0fe6ebc68df894abf873fbd9b93d2cd5c56cbe7b"} Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.199300 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv"] Oct 01 16:13:35 crc kubenswrapper[4726]: W1001 16:13:35.209267 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod965ea363_f11c_4ced_9988_debe220ee3f9.slice/crio-e54d38083cb941df40f0d08c0ff39c62a3ab95407490610c2c2e5324bbfc94f7 WatchSource:0}: Error finding container e54d38083cb941df40f0d08c0ff39c62a3ab95407490610c2c2e5324bbfc94f7: Status 404 returned error can't find the container with id e54d38083cb941df40f0d08c0ff39c62a3ab95407490610c2c2e5324bbfc94f7 Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.210699 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd"] Oct 01 16:13:35 crc kubenswrapper[4726]: E1001 16:13:35.211204 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tnj8l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv_openstack-operators(965ea363-f11c-4ced-9988-debe220ee3f9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:13:35 crc kubenswrapper[4726]: E1001 16:13:35.212341 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv" podUID="965ea363-f11c-4ced-9988-debe220ee3f9" Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.218169 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn"] Oct 01 16:13:35 crc kubenswrapper[4726]: E1001 16:13:35.223309 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nzjkz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-9d6c5db85-z5lfd_openstack-operators(6b63c738-b108-4193-9c9d-11f4eb8227aa): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:13:35 crc kubenswrapper[4726]: E1001 16:13:35.232029 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:acdeebaa51f962066f42f38b6c2d34a62fc6a24f58f9ee63d61b1e0cafbb29f8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r6x4x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-849d5b9b84-w9xkn_openstack-operators(0d89451f-e13a-4195-a2b2-f16443c43463): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.248560 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf"] Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.345040 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r5rkm"] Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.345297 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r5rkm" podUID="6dbcbacf-6b7a-4747-a278-021abf5fd244" containerName="registry-server" containerID="cri-o://f6a73c09add3ddf20d9f298825cf00639945956c574e2f8e8438416311e904f2" gracePeriod=2 Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.430950 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42f6c148-bcc5-4054-88da-dc79d49baeff-cert\") pod \"openstack-operator-controller-manager-5f569c4895-tl4ld\" (UID: \"42f6c148-bcc5-4054-88da-dc79d49baeff\") " pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.436509 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42f6c148-bcc5-4054-88da-dc79d49baeff-cert\") pod \"openstack-operator-controller-manager-5f569c4895-tl4ld\" (UID: \"42f6c148-bcc5-4054-88da-dc79d49baeff\") " pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" Oct 01 16:13:35 crc kubenswrapper[4726]: I1001 16:13:35.722775 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" Oct 01 16:13:35 crc kubenswrapper[4726]: E1001 16:13:35.946354 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f6a73c09add3ddf20d9f298825cf00639945956c574e2f8e8438416311e904f2 is running failed: container process not found" containerID="f6a73c09add3ddf20d9f298825cf00639945956c574e2f8e8438416311e904f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 16:13:35 crc kubenswrapper[4726]: E1001 16:13:35.946734 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f6a73c09add3ddf20d9f298825cf00639945956c574e2f8e8438416311e904f2 is running failed: container process not found" containerID="f6a73c09add3ddf20d9f298825cf00639945956c574e2f8e8438416311e904f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 16:13:35 crc kubenswrapper[4726]: E1001 16:13:35.946941 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f6a73c09add3ddf20d9f298825cf00639945956c574e2f8e8438416311e904f2 is running failed: container process not found" containerID="f6a73c09add3ddf20d9f298825cf00639945956c574e2f8e8438416311e904f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 16:13:35 crc kubenswrapper[4726]: E1001 16:13:35.946967 4726 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f6a73c09add3ddf20d9f298825cf00639945956c574e2f8e8438416311e904f2 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-r5rkm" podUID="6dbcbacf-6b7a-4747-a278-021abf5fd244" containerName="registry-server" Oct 01 16:13:36 crc kubenswrapper[4726]: I1001 16:13:36.144959 4726 generic.go:334] "Generic (PLEG): container finished" podID="6dbcbacf-6b7a-4747-a278-021abf5fd244" containerID="f6a73c09add3ddf20d9f298825cf00639945956c574e2f8e8438416311e904f2" exitCode=0 Oct 01 16:13:36 crc kubenswrapper[4726]: I1001 16:13:36.145286 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5rkm" event={"ID":"6dbcbacf-6b7a-4747-a278-021abf5fd244","Type":"ContainerDied","Data":"f6a73c09add3ddf20d9f298825cf00639945956c574e2f8e8438416311e904f2"} Oct 01 16:13:36 crc kubenswrapper[4726]: I1001 16:13:36.146412 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn" event={"ID":"0d89451f-e13a-4195-a2b2-f16443c43463","Type":"ContainerStarted","Data":"d14fc12a4ab941f3f045443587fb298b2132e945f8036792b0a3de8ff4a82e25"} Oct 01 16:13:36 crc kubenswrapper[4726]: I1001 16:13:36.150132 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv" event={"ID":"965ea363-f11c-4ced-9988-debe220ee3f9","Type":"ContainerStarted","Data":"e54d38083cb941df40f0d08c0ff39c62a3ab95407490610c2c2e5324bbfc94f7"} Oct 01 16:13:36 crc kubenswrapper[4726]: E1001 16:13:36.153741 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv" podUID="965ea363-f11c-4ced-9988-debe220ee3f9" Oct 01 16:13:36 crc kubenswrapper[4726]: I1001 16:13:36.153932 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" event={"ID":"420c55b8-4442-41fa-98a4-338f7338ceb4","Type":"ContainerStarted","Data":"04a92a4f3684175ac6cc7c3985845fdaa0f96733b18c0830a4ebf3b03e6aeaca"} Oct 01 16:13:36 crc kubenswrapper[4726]: I1001 16:13:36.155821 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" event={"ID":"6b63c738-b108-4193-9c9d-11f4eb8227aa","Type":"ContainerStarted","Data":"8aa50ecf047a9e5c4f4e882d60ca20b3382604c13a519d5832610518f7ba548f"} Oct 01 16:13:36 crc kubenswrapper[4726]: I1001 16:13:36.225093 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld"] Oct 01 16:13:37 crc kubenswrapper[4726]: I1001 16:13:37.167140 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" event={"ID":"42f6c148-bcc5-4054-88da-dc79d49baeff","Type":"ContainerStarted","Data":"dc0c46f74ef119c13edbbf9e82aac916c984653874f8453f00856c9bb264843c"} Oct 01 16:13:37 crc kubenswrapper[4726]: E1001 16:13:37.174299 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv" podUID="965ea363-f11c-4ced-9988-debe220ee3f9" Oct 01 16:13:37 crc kubenswrapper[4726]: I1001 16:13:37.568094 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:13:37 crc kubenswrapper[4726]: E1001 16:13:37.594061 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg" podUID="75c9f506-beb2-4fe6-9ada-3ed684131102" Oct 01 16:13:37 crc kubenswrapper[4726]: E1001 16:13:37.656804 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq" podUID="0c8fbaaf-4ce0-4388-855e-fb85534f05e6" Oct 01 16:13:37 crc kubenswrapper[4726]: I1001 16:13:37.664245 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dbcbacf-6b7a-4747-a278-021abf5fd244-utilities\") pod \"6dbcbacf-6b7a-4747-a278-021abf5fd244\" (UID: \"6dbcbacf-6b7a-4747-a278-021abf5fd244\") " Oct 01 16:13:37 crc kubenswrapper[4726]: I1001 16:13:37.664316 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dbcbacf-6b7a-4747-a278-021abf5fd244-catalog-content\") pod \"6dbcbacf-6b7a-4747-a278-021abf5fd244\" (UID: \"6dbcbacf-6b7a-4747-a278-021abf5fd244\") " Oct 01 16:13:37 crc kubenswrapper[4726]: I1001 16:13:37.664358 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26zgg\" (UniqueName: \"kubernetes.io/projected/6dbcbacf-6b7a-4747-a278-021abf5fd244-kube-api-access-26zgg\") pod \"6dbcbacf-6b7a-4747-a278-021abf5fd244\" (UID: \"6dbcbacf-6b7a-4747-a278-021abf5fd244\") " Oct 01 16:13:37 crc kubenswrapper[4726]: I1001 16:13:37.665338 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6dbcbacf-6b7a-4747-a278-021abf5fd244-utilities" (OuterVolumeSpecName: "utilities") pod "6dbcbacf-6b7a-4747-a278-021abf5fd244" (UID: "6dbcbacf-6b7a-4747-a278-021abf5fd244"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:13:37 crc kubenswrapper[4726]: I1001 16:13:37.685269 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dbcbacf-6b7a-4747-a278-021abf5fd244-kube-api-access-26zgg" (OuterVolumeSpecName: "kube-api-access-26zgg") pod "6dbcbacf-6b7a-4747-a278-021abf5fd244" (UID: "6dbcbacf-6b7a-4747-a278-021abf5fd244"). InnerVolumeSpecName "kube-api-access-26zgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:13:37 crc kubenswrapper[4726]: I1001 16:13:37.735200 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6dbcbacf-6b7a-4747-a278-021abf5fd244-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6dbcbacf-6b7a-4747-a278-021abf5fd244" (UID: "6dbcbacf-6b7a-4747-a278-021abf5fd244"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:13:37 crc kubenswrapper[4726]: I1001 16:13:37.766937 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dbcbacf-6b7a-4747-a278-021abf5fd244-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:13:37 crc kubenswrapper[4726]: I1001 16:13:37.766974 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dbcbacf-6b7a-4747-a278-021abf5fd244-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:13:37 crc kubenswrapper[4726]: I1001 16:13:37.766983 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26zgg\" (UniqueName: \"kubernetes.io/projected/6dbcbacf-6b7a-4747-a278-021abf5fd244-kube-api-access-26zgg\") on node \"crc\" DevicePath \"\"" Oct 01 16:13:37 crc kubenswrapper[4726]: E1001 16:13:37.839399 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb" podUID="9f6777bb-5041-4c66-967a-069fa217bf99" Oct 01 16:13:37 crc kubenswrapper[4726]: E1001 16:13:37.856789 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" podUID="6b63c738-b108-4193-9c9d-11f4eb8227aa" Oct 01 16:13:37 crc kubenswrapper[4726]: E1001 16:13:37.857904 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g" podUID="1845e7cd-8784-4994-8f4f-3015bf7d6b3b" Oct 01 16:13:37 crc kubenswrapper[4726]: E1001 16:13:37.858137 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn" podUID="0d89451f-e13a-4195-a2b2-f16443c43463" Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.213233 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq" event={"ID":"0c8fbaaf-4ce0-4388-855e-fb85534f05e6","Type":"ContainerStarted","Data":"e60e4efd73959dcae9ae17ba704f9db0007f7121e5125db34e6054a8b34627e8"} Oct 01 16:13:38 crc kubenswrapper[4726]: E1001 16:13:38.214883 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f\\\"\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq" podUID="0c8fbaaf-4ce0-4388-855e-fb85534f05e6" Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.224230 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" event={"ID":"42f6c148-bcc5-4054-88da-dc79d49baeff","Type":"ContainerStarted","Data":"25068ddacf418232425f8788e7b32497734834ed501931bb4bd6194bf4b7be18"} Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.224277 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" event={"ID":"42f6c148-bcc5-4054-88da-dc79d49baeff","Type":"ContainerStarted","Data":"ad81783d904d180cb14242110ed0f657fa677e1f9649350008b9b91288ded812"} Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.228163 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.242329 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn" event={"ID":"0d89451f-e13a-4195-a2b2-f16443c43463","Type":"ContainerStarted","Data":"8a29c80197420f253482e800470b7fbb79b4bbb1a6fe4d499829f6e10bb7922d"} Oct 01 16:13:38 crc kubenswrapper[4726]: E1001 16:13:38.245547 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:acdeebaa51f962066f42f38b6c2d34a62fc6a24f58f9ee63d61b1e0cafbb29f8\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn" podUID="0d89451f-e13a-4195-a2b2-f16443c43463" Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.246198 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb" event={"ID":"9f6777bb-5041-4c66-967a-069fa217bf99","Type":"ContainerStarted","Data":"3af32365856b542db8b6ad4b7cc6256ec65fe715e79c8d5c7f1ad598aabc2160"} Oct 01 16:13:38 crc kubenswrapper[4726]: E1001 16:13:38.251538 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:15d7b5a365350a831ca59d984df67fadeccf89d599e487a7597b105afb82ce4a\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb" podUID="9f6777bb-5041-4c66-967a-069fa217bf99" Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.261478 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" event={"ID":"6b63c738-b108-4193-9c9d-11f4eb8227aa","Type":"ContainerStarted","Data":"8e6594aa11750d630298fc48697acb21f7105b648750d3e4a8d379a6e83b0458"} Oct 01 16:13:38 crc kubenswrapper[4726]: E1001 16:13:38.273173 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3\\\"\"" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" podUID="6b63c738-b108-4193-9c9d-11f4eb8227aa" Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.276217 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5rkm" event={"ID":"6dbcbacf-6b7a-4747-a278-021abf5fd244","Type":"ContainerDied","Data":"9242e4e8aa443e780eaa9e6dbb93809659849c7dbfd397f7fac9654082a6102c"} Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.276267 4726 scope.go:117] "RemoveContainer" containerID="f6a73c09add3ddf20d9f298825cf00639945956c574e2f8e8438416311e904f2" Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.276378 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5rkm" Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.299700 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg" event={"ID":"75c9f506-beb2-4fe6-9ada-3ed684131102","Type":"ContainerStarted","Data":"d8cd515299b17ace39d97889f2614db9a5ec1de07ab35bc34db5eafe8d2b82b4"} Oct 01 16:13:38 crc kubenswrapper[4726]: E1001 16:13:38.312676 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/openstack-k8s-operators/telemetry-operator:020bd70758b454f4f03f761350467eb9337e2b6e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg" podUID="75c9f506-beb2-4fe6-9ada-3ed684131102" Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.315721 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g" event={"ID":"1845e7cd-8784-4994-8f4f-3015bf7d6b3b","Type":"ContainerStarted","Data":"51cbf121e0d11de7cd119a8beda4a153e79d2f38be5e55c5cdd433d18ea24da4"} Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.333166 4726 scope.go:117] "RemoveContainer" containerID="f542bbbfa187bb76d9aed6b04285cb8b93acf8bcb87f7a945c41e00927ccd02d" Oct 01 16:13:38 crc kubenswrapper[4726]: E1001 16:13:38.333267 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g" podUID="1845e7cd-8784-4994-8f4f-3015bf7d6b3b" Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.359566 4726 scope.go:117] "RemoveContainer" containerID="3c9278e0b3673e86645af6f1c2cf8b4a60871ce7282811caaff41ad16f807cf9" Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.521435 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r5rkm"] Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.526944 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r5rkm"] Oct 01 16:13:38 crc kubenswrapper[4726]: I1001 16:13:38.632293 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" podStartSLOduration=5.632277442 podStartE2EDuration="5.632277442s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:13:38.626393959 +0000 UTC m=+871.527946536" watchObservedRunningTime="2025-10-01 16:13:38.632277442 +0000 UTC m=+871.533830019" Oct 01 16:13:39 crc kubenswrapper[4726]: E1001 16:13:39.329950 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.163:5001/openstack-k8s-operators/telemetry-operator:020bd70758b454f4f03f761350467eb9337e2b6e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg" podUID="75c9f506-beb2-4fe6-9ada-3ed684131102" Oct 01 16:13:39 crc kubenswrapper[4726]: E1001 16:13:39.334578 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g" podUID="1845e7cd-8784-4994-8f4f-3015bf7d6b3b" Oct 01 16:13:39 crc kubenswrapper[4726]: E1001 16:13:39.334751 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:15d7b5a365350a831ca59d984df67fadeccf89d599e487a7597b105afb82ce4a\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb" podUID="9f6777bb-5041-4c66-967a-069fa217bf99" Oct 01 16:13:39 crc kubenswrapper[4726]: E1001 16:13:39.334739 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3\\\"\"" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" podUID="6b63c738-b108-4193-9c9d-11f4eb8227aa" Oct 01 16:13:39 crc kubenswrapper[4726]: E1001 16:13:39.334971 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:acdeebaa51f962066f42f38b6c2d34a62fc6a24f58f9ee63d61b1e0cafbb29f8\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn" podUID="0d89451f-e13a-4195-a2b2-f16443c43463" Oct 01 16:13:39 crc kubenswrapper[4726]: E1001 16:13:39.342166 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f\\\"\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq" podUID="0c8fbaaf-4ce0-4388-855e-fb85534f05e6" Oct 01 16:13:39 crc kubenswrapper[4726]: I1001 16:13:39.816348 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dbcbacf-6b7a-4747-a278-021abf5fd244" path="/var/lib/kubelet/pods/6dbcbacf-6b7a-4747-a278-021abf5fd244/volumes" Oct 01 16:13:45 crc kubenswrapper[4726]: I1001 16:13:45.729337 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5f569c4895-tl4ld" Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.425289 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-nv4ww" event={"ID":"0396f184-2150-4944-a81c-c0d0fafe1317","Type":"ContainerStarted","Data":"fdcadc37c62e672201186bcdf321ad75319a684292f8ab21a3c945a36874f67d"} Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.455660 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-zxd4h" event={"ID":"117d55f2-eee8-454b-9091-3ffffd61c547","Type":"ContainerStarted","Data":"ed865b21ef0ace0e5fe0422303ab15e8dfe1f4d2245039ac71f4bf7d6003285d"} Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.465043 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-cpbbq" event={"ID":"8e085c54-533c-4186-a6fe-5d8a5ccbac2f","Type":"ContainerStarted","Data":"43e746655020435b0652b4d75c1150db9ad53acf0266dcf2f9216b5888d5b061"} Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.486728 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-wgkz2" event={"ID":"71d9bfcf-89ac-440f-b2b1-1c2969859b0d","Type":"ContainerStarted","Data":"6f5e9f059a0a08e6e28375e39fd9dc7ab8df4e87eec2c900ba983e9f06e70783"} Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.500572 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-t2vwz" event={"ID":"b9f1f46c-c6fa-443f-b53d-54614779c0cc","Type":"ContainerStarted","Data":"d3d15994f135ebfee9eac8043c720ea6de629738c7360c0ea471922e5d3a70f1"} Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.510697 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-qlb92" event={"ID":"f82a5172-6387-4a24-a7e4-a5bd4c90d310","Type":"ContainerStarted","Data":"f28483bf1c23e12d4f1971c14cbf29258ca2e6469ef954769c3f46fdb4eb1152"} Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.524959 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-5hvl7" event={"ID":"23922646-bd87-424b-85d7-8a2a55056438","Type":"ContainerStarted","Data":"e17836fd6045cc93fbdced1fa22945df6f07b2f678b1277ec26745768722b48a"} Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.540676 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-qd8pc" event={"ID":"2d90550d-aedc-4e50-81f9-da6f285f8c2b","Type":"ContainerStarted","Data":"8d4df7aeb71e2017e40d0920b3554798e37e724098388ff628576d3758eab981"} Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.574317 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9nqm2" event={"ID":"b824be4d-8e37-42b5-8fe2-0dc6a1c4b4a1","Type":"ContainerStarted","Data":"60767a0456397c401edbb112b9d99d5a20cc964662c2080dfa7036c9b062dcb1"} Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.654933 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-67rj9" event={"ID":"daf7d005-ec55-4cdb-91d6-eac465e42e17","Type":"ContainerStarted","Data":"0fc272ed25456c73165cdcab15a6ca0ae41d250b6945cb67e17872af8748771c"} Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.664285 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-7s6mh" event={"ID":"ff753ea2-fa5b-4367-b476-596fd49d0557","Type":"ContainerStarted","Data":"30e9a93fbf165b553781ef113f70528f02f673fd36db86d3f081e92a1a0ee32d"} Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.709173 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wckhv" event={"ID":"85f77f5d-6b74-40b5-b238-d3515b4a685c","Type":"ContainerStarted","Data":"3d180cabb3277e8248660dfc16199dfe6bd809b00929b3e2d9c9a0ecc8f157b7"} Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.709247 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wckhv" Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.755401 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-pl7jv" event={"ID":"fa99155f-85e2-406e-84c4-4a95ab74407f","Type":"ContainerStarted","Data":"a6947cc5ea2406ddffdca32f4136dfc1dcdc1367e162896fcb2e55778d021546"} Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.756443 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-pl7jv" Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.756511 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wckhv" podStartSLOduration=3.5868008 podStartE2EDuration="16.756495616s" podCreationTimestamp="2025-10-01 16:13:32 +0000 UTC" firstStartedPulling="2025-10-01 16:13:34.464134417 +0000 UTC m=+867.365686994" lastFinishedPulling="2025-10-01 16:13:47.633829233 +0000 UTC m=+880.535381810" observedRunningTime="2025-10-01 16:13:48.753490458 +0000 UTC m=+881.655043035" watchObservedRunningTime="2025-10-01 16:13:48.756495616 +0000 UTC m=+881.658048193" Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.779788 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" event={"ID":"420c55b8-4442-41fa-98a4-338f7338ceb4","Type":"ContainerStarted","Data":"843d97cdc8571ce5203989843e8db1e758404c35ece302e292ebf7c0e1aad89b"} Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.782143 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-pl7jv" podStartSLOduration=3.171603767 podStartE2EDuration="15.7821251s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.050271542 +0000 UTC m=+867.951824119" lastFinishedPulling="2025-10-01 16:13:47.660792865 +0000 UTC m=+880.562345452" observedRunningTime="2025-10-01 16:13:48.780200463 +0000 UTC m=+881.681753040" watchObservedRunningTime="2025-10-01 16:13:48.7821251 +0000 UTC m=+881.683678017" Oct 01 16:13:48 crc kubenswrapper[4726]: I1001 16:13:48.793915 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zjwvl" event={"ID":"39860f7b-9a7f-44b1-8bf4-7a2204092c18","Type":"ContainerStarted","Data":"ba285885e9395423bcec420b6411a4d98e3fc42aefd91eb86e01d99a824a46ae"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.802392 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-67rj9" event={"ID":"daf7d005-ec55-4cdb-91d6-eac465e42e17","Type":"ContainerStarted","Data":"de7baf08cc2247b3fb62933b05bdd0c815cb4a60a8ab91ba37f6b1ba0303d524"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.802745 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-67rj9" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.804231 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-7s6mh" event={"ID":"ff753ea2-fa5b-4367-b476-596fd49d0557","Type":"ContainerStarted","Data":"1a43425cd16c777f02d13216234a5a0c7254c1783cf99393a96234755a4a4bc1"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.804489 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-7s6mh" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.806165 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-t2vwz" event={"ID":"b9f1f46c-c6fa-443f-b53d-54614779c0cc","Type":"ContainerStarted","Data":"6fda988af9903db1d158f0c3b31d414fce6fcaca8e2c8ca7a7edb8b87a480420"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.826458 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-67rj9" podStartSLOduration=4.571888111 podStartE2EDuration="17.826438441s" podCreationTimestamp="2025-10-01 16:13:32 +0000 UTC" firstStartedPulling="2025-10-01 16:13:34.455215924 +0000 UTC m=+867.356768501" lastFinishedPulling="2025-10-01 16:13:47.709766234 +0000 UTC m=+880.611318831" observedRunningTime="2025-10-01 16:13:49.821490625 +0000 UTC m=+882.723043212" watchObservedRunningTime="2025-10-01 16:13:49.826438441 +0000 UTC m=+882.727991018" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.831687 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-t2vwz" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.831721 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-cpbbq" event={"ID":"8e085c54-533c-4186-a6fe-5d8a5ccbac2f","Type":"ContainerStarted","Data":"685f41c617ee25843043b3b7abc86c4e0bb6615ae32638e16f86d144777a13c7"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.831750 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-cpbbq" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.831765 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9nqm2" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.831779 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-85777745bb-nv4ww" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.831790 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9nqm2" event={"ID":"b824be4d-8e37-42b5-8fe2-0dc6a1c4b4a1","Type":"ContainerStarted","Data":"630b19515f8f583794f814559dc793d66922a5a1ed291ad2da19c33db193a02b"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.831803 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-nv4ww" event={"ID":"0396f184-2150-4944-a81c-c0d0fafe1317","Type":"ContainerStarted","Data":"117892b2f60977be5242bbd77f0c9c9fa57195d5a28ea587223c7c07ddc97687"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.855191 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9nqm2" podStartSLOduration=4.219520254 podStartE2EDuration="16.855170625s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.047598694 +0000 UTC m=+867.949151271" lastFinishedPulling="2025-10-01 16:13:47.683249025 +0000 UTC m=+880.584801642" observedRunningTime="2025-10-01 16:13:49.850931051 +0000 UTC m=+882.752483638" watchObservedRunningTime="2025-10-01 16:13:49.855170625 +0000 UTC m=+882.756723202" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.856095 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" event={"ID":"420c55b8-4442-41fa-98a4-338f7338ceb4","Type":"ContainerStarted","Data":"3590eb55f67f1c8e8f0715696bb2770b3f6706e517118046b6bf046c23e78ccd"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.860434 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.877831 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zjwvl" event={"ID":"39860f7b-9a7f-44b1-8bf4-7a2204092c18","Type":"ContainerStarted","Data":"2f1e0ca905c5803f1a2563a2e9a9af1f246eaff9c84bc5810790e9328ae9b91e"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.877894 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zjwvl" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.884318 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-t2vwz" podStartSLOduration=4.305183281 podStartE2EDuration="16.884297711s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.074535705 +0000 UTC m=+867.976088282" lastFinishedPulling="2025-10-01 16:13:47.653650115 +0000 UTC m=+880.555202712" observedRunningTime="2025-10-01 16:13:49.876414139 +0000 UTC m=+882.777966726" watchObservedRunningTime="2025-10-01 16:13:49.884297711 +0000 UTC m=+882.785850288" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.892898 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-qlb92" event={"ID":"f82a5172-6387-4a24-a7e4-a5bd4c90d310","Type":"ContainerStarted","Data":"f9ffb530c04a739c31271bb23bae1cc1835401b607f8a116297d19a536470433"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.893445 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-qlb92" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.894670 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-zxd4h" event={"ID":"117d55f2-eee8-454b-9091-3ffffd61c547","Type":"ContainerStarted","Data":"ebe75abe59190c2d0ee9e76978e77be40a4d58fa3979526574dbb7ebbd07ec94"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.895354 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-zxd4h" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.900904 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-cpbbq" podStartSLOduration=4.292311474 podStartE2EDuration="16.900881939s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.076453232 +0000 UTC m=+867.978005809" lastFinishedPulling="2025-10-01 16:13:47.685023657 +0000 UTC m=+880.586576274" observedRunningTime="2025-10-01 16:13:49.896275853 +0000 UTC m=+882.797828430" watchObservedRunningTime="2025-10-01 16:13:49.900881939 +0000 UTC m=+882.802434516" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.902609 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-5hvl7" event={"ID":"23922646-bd87-424b-85d7-8a2a55056438","Type":"ContainerStarted","Data":"51f459d7420b0c205a4b6f1a3c4a7a974208498fc27f49fd0c802b7f01640720"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.903436 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-5hvl7" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.908018 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-pl7jv" event={"ID":"fa99155f-85e2-406e-84c4-4a95ab74407f","Type":"ContainerStarted","Data":"d0c53111206bfd9735c65c4dea96c51f7643d0128afd2d568e200a975dc89568"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.910227 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-wgkz2" event={"ID":"71d9bfcf-89ac-440f-b2b1-1c2969859b0d","Type":"ContainerStarted","Data":"ee455383475a975798290acffcac7807a581f72d2e827e55c9060c460f9d240e"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.910828 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-wgkz2" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.920436 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wckhv" event={"ID":"85f77f5d-6b74-40b5-b238-d3515b4a685c","Type":"ContainerStarted","Data":"198e7993c22708a3e99c029899cc1931b550d41ea15c82bf7ef3d5e6e94470c9"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.921953 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-qd8pc" event={"ID":"2d90550d-aedc-4e50-81f9-da6f285f8c2b","Type":"ContainerStarted","Data":"7642d6d84a4a833cabd45fd5b45d02ecf1c4f3848a0dffad4139b7f92c0966db"} Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.922413 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-qd8pc" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.931526 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-7s6mh" podStartSLOduration=4.322305085 podStartE2EDuration="16.931508979s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.070736544 +0000 UTC m=+867.972289121" lastFinishedPulling="2025-10-01 16:13:47.679940398 +0000 UTC m=+880.581493015" observedRunningTime="2025-10-01 16:13:49.910976325 +0000 UTC m=+882.812528902" watchObservedRunningTime="2025-10-01 16:13:49.931508979 +0000 UTC m=+882.833061556" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.937415 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-5hvl7" podStartSLOduration=4.31396462 podStartE2EDuration="16.937398402s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.054201478 +0000 UTC m=+867.955754055" lastFinishedPulling="2025-10-01 16:13:47.67763522 +0000 UTC m=+880.579187837" observedRunningTime="2025-10-01 16:13:49.934736033 +0000 UTC m=+882.836288610" watchObservedRunningTime="2025-10-01 16:13:49.937398402 +0000 UTC m=+882.838950979" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.956721 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zjwvl" podStartSLOduration=4.9695558680000005 podStartE2EDuration="17.956700849s" podCreationTimestamp="2025-10-01 16:13:32 +0000 UTC" firstStartedPulling="2025-10-01 16:13:34.674812138 +0000 UTC m=+867.576364725" lastFinishedPulling="2025-10-01 16:13:47.661957119 +0000 UTC m=+880.563509706" observedRunningTime="2025-10-01 16:13:49.950404964 +0000 UTC m=+882.851957551" watchObservedRunningTime="2025-10-01 16:13:49.956700849 +0000 UTC m=+882.858253426" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.968737 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-zxd4h" podStartSLOduration=4.614537884 podStartE2EDuration="17.968718872s" podCreationTimestamp="2025-10-01 16:13:32 +0000 UTC" firstStartedPulling="2025-10-01 16:13:34.261614015 +0000 UTC m=+867.163166582" lastFinishedPulling="2025-10-01 16:13:47.615794993 +0000 UTC m=+880.517347570" observedRunningTime="2025-10-01 16:13:49.965480807 +0000 UTC m=+882.867033394" watchObservedRunningTime="2025-10-01 16:13:49.968718872 +0000 UTC m=+882.870271459" Oct 01 16:13:49 crc kubenswrapper[4726]: I1001 16:13:49.984147 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-wgkz2" podStartSLOduration=4.733585453 podStartE2EDuration="17.984130795s" podCreationTimestamp="2025-10-01 16:13:32 +0000 UTC" firstStartedPulling="2025-10-01 16:13:34.434484815 +0000 UTC m=+867.336037392" lastFinishedPulling="2025-10-01 16:13:47.685030147 +0000 UTC m=+880.586582734" observedRunningTime="2025-10-01 16:13:49.98226621 +0000 UTC m=+882.883818807" watchObservedRunningTime="2025-10-01 16:13:49.984130795 +0000 UTC m=+882.885683362" Oct 01 16:13:50 crc kubenswrapper[4726]: I1001 16:13:50.012419 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" podStartSLOduration=4.613727269 podStartE2EDuration="17.012397856s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.27919083 +0000 UTC m=+868.180743407" lastFinishedPulling="2025-10-01 16:13:47.677861407 +0000 UTC m=+880.579413994" observedRunningTime="2025-10-01 16:13:50.005092701 +0000 UTC m=+882.906645288" watchObservedRunningTime="2025-10-01 16:13:50.012397856 +0000 UTC m=+882.913950443" Oct 01 16:13:50 crc kubenswrapper[4726]: I1001 16:13:50.029085 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-qlb92" podStartSLOduration=5.063869969 podStartE2EDuration="18.029064775s" podCreationTimestamp="2025-10-01 16:13:32 +0000 UTC" firstStartedPulling="2025-10-01 16:13:34.713792024 +0000 UTC m=+867.615344601" lastFinishedPulling="2025-10-01 16:13:47.67898682 +0000 UTC m=+880.580539407" observedRunningTime="2025-10-01 16:13:50.02480682 +0000 UTC m=+882.926359407" watchObservedRunningTime="2025-10-01 16:13:50.029064775 +0000 UTC m=+882.930617352" Oct 01 16:13:50 crc kubenswrapper[4726]: I1001 16:13:50.040769 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-85777745bb-nv4ww" podStartSLOduration=4.399750639 podStartE2EDuration="17.040754188s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.04271495 +0000 UTC m=+867.944267527" lastFinishedPulling="2025-10-01 16:13:47.683718489 +0000 UTC m=+880.585271076" observedRunningTime="2025-10-01 16:13:50.039776979 +0000 UTC m=+882.941329566" watchObservedRunningTime="2025-10-01 16:13:50.040754188 +0000 UTC m=+882.942306765" Oct 01 16:13:50 crc kubenswrapper[4726]: I1001 16:13:50.070529 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-qd8pc" podStartSLOduration=4.399783921 podStartE2EDuration="17.070508473s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.00322166 +0000 UTC m=+867.904774237" lastFinishedPulling="2025-10-01 16:13:47.673946202 +0000 UTC m=+880.575498789" observedRunningTime="2025-10-01 16:13:50.066491955 +0000 UTC m=+882.968044542" watchObservedRunningTime="2025-10-01 16:13:50.070508473 +0000 UTC m=+882.972061050" Oct 01 16:13:51 crc kubenswrapper[4726]: I1001 16:13:51.941901 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv" event={"ID":"965ea363-f11c-4ced-9988-debe220ee3f9","Type":"ContainerStarted","Data":"4d3bcda6d4866a9296f6a4b4a9460af25b786bf8898084d7b795319c886f374e"} Oct 01 16:13:51 crc kubenswrapper[4726]: I1001 16:13:51.963852 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv" podStartSLOduration=2.9078712859999998 podStartE2EDuration="18.963833035s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.211083568 +0000 UTC m=+868.112636145" lastFinishedPulling="2025-10-01 16:13:51.267045317 +0000 UTC m=+884.168597894" observedRunningTime="2025-10-01 16:13:51.958751025 +0000 UTC m=+884.860303652" watchObservedRunningTime="2025-10-01 16:13:51.963833035 +0000 UTC m=+884.865385602" Oct 01 16:13:53 crc kubenswrapper[4726]: I1001 16:13:53.226241 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-zxd4h" Oct 01 16:13:53 crc kubenswrapper[4726]: I1001 16:13:53.244690 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wckhv" Oct 01 16:13:53 crc kubenswrapper[4726]: I1001 16:13:53.294205 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-wgkz2" Oct 01 16:13:53 crc kubenswrapper[4726]: I1001 16:13:53.317631 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-67rj9" Oct 01 16:13:53 crc kubenswrapper[4726]: I1001 16:13:53.370428 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-qlb92" Oct 01 16:13:53 crc kubenswrapper[4726]: I1001 16:13:53.439646 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-zjwvl" Oct 01 16:13:53 crc kubenswrapper[4726]: I1001 16:13:53.482006 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-cpbbq" Oct 01 16:13:53 crc kubenswrapper[4726]: I1001 16:13:53.556316 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-qd8pc" Oct 01 16:13:53 crc kubenswrapper[4726]: I1001 16:13:53.591730 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-5hvl7" Oct 01 16:13:53 crc kubenswrapper[4726]: I1001 16:13:53.651892 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-t2vwz" Oct 01 16:13:53 crc kubenswrapper[4726]: I1001 16:13:53.674803 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-7s6mh" Oct 01 16:13:53 crc kubenswrapper[4726]: I1001 16:13:53.777797 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-pl7jv" Oct 01 16:13:53 crc kubenswrapper[4726]: I1001 16:13:53.883161 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-85777745bb-nv4ww" Oct 01 16:13:53 crc kubenswrapper[4726]: I1001 16:13:53.883260 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-9nqm2" Oct 01 16:13:54 crc kubenswrapper[4726]: I1001 16:13:54.345845 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf" Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.027367 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" event={"ID":"6b63c738-b108-4193-9c9d-11f4eb8227aa","Type":"ContainerStarted","Data":"05c4d563b3a4362d320d4dfc749f46713f582c2c106c140222c428f387cf3a93"} Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.027966 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.030963 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg" event={"ID":"75c9f506-beb2-4fe6-9ada-3ed684131102","Type":"ContainerStarted","Data":"b27b09e4c742e0bcf13053b62cf284d90ed3e0530e1ba2a379dc9f0e8198dc34"} Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.031188 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg" Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.042765 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g" event={"ID":"1845e7cd-8784-4994-8f4f-3015bf7d6b3b","Type":"ContainerStarted","Data":"f24dc2b726ee97f9fe69ec0755bea4baa29766e251e2d7c32faf99fefa4809a9"} Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.043034 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g" Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.047196 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq" event={"ID":"0c8fbaaf-4ce0-4388-855e-fb85534f05e6","Type":"ContainerStarted","Data":"a18824f0c8b570f34d52217f9ea2d31473ccc6c058eb9cac1327306d32ff87e9"} Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.047892 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq" Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.052803 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" podStartSLOduration=3.343917361 podStartE2EDuration="26.052786598s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.22303538 +0000 UTC m=+868.124587957" lastFinishedPulling="2025-10-01 16:13:57.931904577 +0000 UTC m=+890.833457194" observedRunningTime="2025-10-01 16:13:59.048230714 +0000 UTC m=+891.949783301" watchObservedRunningTime="2025-10-01 16:13:59.052786598 +0000 UTC m=+891.954339175" Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.056263 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn" event={"ID":"0d89451f-e13a-4195-a2b2-f16443c43463","Type":"ContainerStarted","Data":"5112f962f0b0f1b89b7d2771043f9c2e34536ec975f465f33d984a47ceee67a9"} Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.056550 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn" Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.059254 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb" event={"ID":"9f6777bb-5041-4c66-967a-069fa217bf99","Type":"ContainerStarted","Data":"06f9b7d1434b850dbd96b72079532d094448f9171bf9784d8857d1b9bf4f7ec4"} Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.059490 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb" Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.070769 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq" podStartSLOduration=3.253989668 podStartE2EDuration="26.070755446s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.083578621 +0000 UTC m=+867.985131198" lastFinishedPulling="2025-10-01 16:13:57.900344399 +0000 UTC m=+890.801896976" observedRunningTime="2025-10-01 16:13:59.065746689 +0000 UTC m=+891.967299286" watchObservedRunningTime="2025-10-01 16:13:59.070755446 +0000 UTC m=+891.972308023" Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.081455 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg" podStartSLOduration=3.7188356689999997 podStartE2EDuration="26.08143622s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.078954245 +0000 UTC m=+867.980506832" lastFinishedPulling="2025-10-01 16:13:57.441554766 +0000 UTC m=+890.343107383" observedRunningTime="2025-10-01 16:13:59.079178274 +0000 UTC m=+891.980730861" watchObservedRunningTime="2025-10-01 16:13:59.08143622 +0000 UTC m=+891.982988797" Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.098987 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g" podStartSLOduration=3.277486858 podStartE2EDuration="26.098966645s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.078920124 +0000 UTC m=+867.980472701" lastFinishedPulling="2025-10-01 16:13:57.900399911 +0000 UTC m=+890.801952488" observedRunningTime="2025-10-01 16:13:59.093513965 +0000 UTC m=+891.995066552" watchObservedRunningTime="2025-10-01 16:13:59.098966645 +0000 UTC m=+892.000519222" Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.112117 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb" podStartSLOduration=3.757964699 podStartE2EDuration="26.112098111s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.087409304 +0000 UTC m=+867.988961881" lastFinishedPulling="2025-10-01 16:13:57.441542706 +0000 UTC m=+890.343095293" observedRunningTime="2025-10-01 16:13:59.110861605 +0000 UTC m=+892.012414192" watchObservedRunningTime="2025-10-01 16:13:59.112098111 +0000 UTC m=+892.013650688" Oct 01 16:13:59 crc kubenswrapper[4726]: I1001 16:13:59.141994 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn" podStartSLOduration=3.411000772 podStartE2EDuration="26.141976599s" podCreationTimestamp="2025-10-01 16:13:33 +0000 UTC" firstStartedPulling="2025-10-01 16:13:35.231827238 +0000 UTC m=+868.133379815" lastFinishedPulling="2025-10-01 16:13:57.962803035 +0000 UTC m=+890.864355642" observedRunningTime="2025-10-01 16:13:59.134938303 +0000 UTC m=+892.036490890" watchObservedRunningTime="2025-10-01 16:13:59.141976599 +0000 UTC m=+892.043529176" Oct 01 16:14:03 crc kubenswrapper[4726]: I1001 16:14:03.614627 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-5vlmb" Oct 01 16:14:03 crc kubenswrapper[4726]: I1001 16:14:03.636784 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-w9xkn" Oct 01 16:14:03 crc kubenswrapper[4726]: I1001 16:14:03.640662 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-zdkvq" Oct 01 16:14:03 crc kubenswrapper[4726]: I1001 16:14:03.732385 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-j4t5g" Oct 01 16:14:03 crc kubenswrapper[4726]: I1001 16:14:03.837332 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7866c8b47c-z6rlg" Oct 01 16:14:04 crc kubenswrapper[4726]: I1001 16:14:04.074821 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-z5lfd" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.694943 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mxntn"] Oct 01 16:14:19 crc kubenswrapper[4726]: E1001 16:14:19.699949 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dbcbacf-6b7a-4747-a278-021abf5fd244" containerName="extract-content" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.700066 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dbcbacf-6b7a-4747-a278-021abf5fd244" containerName="extract-content" Oct 01 16:14:19 crc kubenswrapper[4726]: E1001 16:14:19.700154 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dbcbacf-6b7a-4747-a278-021abf5fd244" containerName="extract-utilities" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.700215 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dbcbacf-6b7a-4747-a278-021abf5fd244" containerName="extract-utilities" Oct 01 16:14:19 crc kubenswrapper[4726]: E1001 16:14:19.700286 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dbcbacf-6b7a-4747-a278-021abf5fd244" containerName="registry-server" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.700338 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dbcbacf-6b7a-4747-a278-021abf5fd244" containerName="registry-server" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.700549 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dbcbacf-6b7a-4747-a278-021abf5fd244" containerName="registry-server" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.701269 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mxntn"] Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.701413 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mxntn" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.703781 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.704177 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-r79fs" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.706163 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.706188 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.766879 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rrn75"] Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.768269 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.783220 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.789501 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add42128-3a49-40ae-baec-e74f857203fc-config\") pod \"dnsmasq-dns-675f4bcbfc-mxntn\" (UID: \"add42128-3a49-40ae-baec-e74f857203fc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mxntn" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.789547 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tls67\" (UniqueName: \"kubernetes.io/projected/add42128-3a49-40ae-baec-e74f857203fc-kube-api-access-tls67\") pod \"dnsmasq-dns-675f4bcbfc-mxntn\" (UID: \"add42128-3a49-40ae-baec-e74f857203fc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mxntn" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.789712 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rrn75"] Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.890345 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tls67\" (UniqueName: \"kubernetes.io/projected/add42128-3a49-40ae-baec-e74f857203fc-kube-api-access-tls67\") pod \"dnsmasq-dns-675f4bcbfc-mxntn\" (UID: \"add42128-3a49-40ae-baec-e74f857203fc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mxntn" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.892137 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv4jw\" (UniqueName: \"kubernetes.io/projected/0d7affd8-5a89-4ec3-a593-b4dfe592861f-kube-api-access-vv4jw\") pod \"dnsmasq-dns-78dd6ddcc-rrn75\" (UID: \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.892204 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d7affd8-5a89-4ec3-a593-b4dfe592861f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rrn75\" (UID: \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.892232 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d7affd8-5a89-4ec3-a593-b4dfe592861f-config\") pod \"dnsmasq-dns-78dd6ddcc-rrn75\" (UID: \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.892307 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add42128-3a49-40ae-baec-e74f857203fc-config\") pod \"dnsmasq-dns-675f4bcbfc-mxntn\" (UID: \"add42128-3a49-40ae-baec-e74f857203fc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mxntn" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.893418 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add42128-3a49-40ae-baec-e74f857203fc-config\") pod \"dnsmasq-dns-675f4bcbfc-mxntn\" (UID: \"add42128-3a49-40ae-baec-e74f857203fc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mxntn" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.909062 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tls67\" (UniqueName: \"kubernetes.io/projected/add42128-3a49-40ae-baec-e74f857203fc-kube-api-access-tls67\") pod \"dnsmasq-dns-675f4bcbfc-mxntn\" (UID: \"add42128-3a49-40ae-baec-e74f857203fc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mxntn" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.992930 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d7affd8-5a89-4ec3-a593-b4dfe592861f-config\") pod \"dnsmasq-dns-78dd6ddcc-rrn75\" (UID: \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.993033 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv4jw\" (UniqueName: \"kubernetes.io/projected/0d7affd8-5a89-4ec3-a593-b4dfe592861f-kube-api-access-vv4jw\") pod \"dnsmasq-dns-78dd6ddcc-rrn75\" (UID: \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.993128 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d7affd8-5a89-4ec3-a593-b4dfe592861f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rrn75\" (UID: \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.993841 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d7affd8-5a89-4ec3-a593-b4dfe592861f-config\") pod \"dnsmasq-dns-78dd6ddcc-rrn75\" (UID: \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" Oct 01 16:14:19 crc kubenswrapper[4726]: I1001 16:14:19.993870 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d7affd8-5a89-4ec3-a593-b4dfe592861f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rrn75\" (UID: \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" Oct 01 16:14:20 crc kubenswrapper[4726]: I1001 16:14:20.015187 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv4jw\" (UniqueName: \"kubernetes.io/projected/0d7affd8-5a89-4ec3-a593-b4dfe592861f-kube-api-access-vv4jw\") pod \"dnsmasq-dns-78dd6ddcc-rrn75\" (UID: \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" Oct 01 16:14:20 crc kubenswrapper[4726]: I1001 16:14:20.016973 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mxntn" Oct 01 16:14:20 crc kubenswrapper[4726]: I1001 16:14:20.094312 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" Oct 01 16:14:20 crc kubenswrapper[4726]: I1001 16:14:20.429440 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mxntn"] Oct 01 16:14:20 crc kubenswrapper[4726]: W1001 16:14:20.430935 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadd42128_3a49_40ae_baec_e74f857203fc.slice/crio-58b75bccfa69b77129cdf6900ad4cd2f2a432bacfeb4b8b2397d8b43e861508e WatchSource:0}: Error finding container 58b75bccfa69b77129cdf6900ad4cd2f2a432bacfeb4b8b2397d8b43e861508e: Status 404 returned error can't find the container with id 58b75bccfa69b77129cdf6900ad4cd2f2a432bacfeb4b8b2397d8b43e861508e Oct 01 16:14:20 crc kubenswrapper[4726]: I1001 16:14:20.553904 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rrn75"] Oct 01 16:14:20 crc kubenswrapper[4726]: W1001 16:14:20.554581 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d7affd8_5a89_4ec3_a593_b4dfe592861f.slice/crio-acac541f6a01907ab4cb6040e09ec89c22435e64d40428e61af76478e10ef95b WatchSource:0}: Error finding container acac541f6a01907ab4cb6040e09ec89c22435e64d40428e61af76478e10ef95b: Status 404 returned error can't find the container with id acac541f6a01907ab4cb6040e09ec89c22435e64d40428e61af76478e10ef95b Oct 01 16:14:21 crc kubenswrapper[4726]: I1001 16:14:21.247309 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" event={"ID":"0d7affd8-5a89-4ec3-a593-b4dfe592861f","Type":"ContainerStarted","Data":"acac541f6a01907ab4cb6040e09ec89c22435e64d40428e61af76478e10ef95b"} Oct 01 16:14:21 crc kubenswrapper[4726]: I1001 16:14:21.254249 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-mxntn" event={"ID":"add42128-3a49-40ae-baec-e74f857203fc","Type":"ContainerStarted","Data":"58b75bccfa69b77129cdf6900ad4cd2f2a432bacfeb4b8b2397d8b43e861508e"} Oct 01 16:14:22 crc kubenswrapper[4726]: I1001 16:14:22.933610 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mxntn"] Oct 01 16:14:22 crc kubenswrapper[4726]: I1001 16:14:22.965162 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qsbql"] Oct 01 16:14:22 crc kubenswrapper[4726]: I1001 16:14:22.966411 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:14:22 crc kubenswrapper[4726]: I1001 16:14:22.981640 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qsbql"] Oct 01 16:14:22 crc kubenswrapper[4726]: I1001 16:14:22.983243 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47538946-adc6-42ea-a62a-3cf2fb6a3206-config\") pod \"dnsmasq-dns-666b6646f7-qsbql\" (UID: \"47538946-adc6-42ea-a62a-3cf2fb6a3206\") " pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:14:22 crc kubenswrapper[4726]: I1001 16:14:22.983296 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5spq9\" (UniqueName: \"kubernetes.io/projected/47538946-adc6-42ea-a62a-3cf2fb6a3206-kube-api-access-5spq9\") pod \"dnsmasq-dns-666b6646f7-qsbql\" (UID: \"47538946-adc6-42ea-a62a-3cf2fb6a3206\") " pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:14:22 crc kubenswrapper[4726]: I1001 16:14:22.983356 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47538946-adc6-42ea-a62a-3cf2fb6a3206-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qsbql\" (UID: \"47538946-adc6-42ea-a62a-3cf2fb6a3206\") " pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.084696 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47538946-adc6-42ea-a62a-3cf2fb6a3206-config\") pod \"dnsmasq-dns-666b6646f7-qsbql\" (UID: \"47538946-adc6-42ea-a62a-3cf2fb6a3206\") " pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.084747 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5spq9\" (UniqueName: \"kubernetes.io/projected/47538946-adc6-42ea-a62a-3cf2fb6a3206-kube-api-access-5spq9\") pod \"dnsmasq-dns-666b6646f7-qsbql\" (UID: \"47538946-adc6-42ea-a62a-3cf2fb6a3206\") " pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.084790 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47538946-adc6-42ea-a62a-3cf2fb6a3206-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qsbql\" (UID: \"47538946-adc6-42ea-a62a-3cf2fb6a3206\") " pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.085687 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47538946-adc6-42ea-a62a-3cf2fb6a3206-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qsbql\" (UID: \"47538946-adc6-42ea-a62a-3cf2fb6a3206\") " pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.086215 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47538946-adc6-42ea-a62a-3cf2fb6a3206-config\") pod \"dnsmasq-dns-666b6646f7-qsbql\" (UID: \"47538946-adc6-42ea-a62a-3cf2fb6a3206\") " pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.109129 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5spq9\" (UniqueName: \"kubernetes.io/projected/47538946-adc6-42ea-a62a-3cf2fb6a3206-kube-api-access-5spq9\") pod \"dnsmasq-dns-666b6646f7-qsbql\" (UID: \"47538946-adc6-42ea-a62a-3cf2fb6a3206\") " pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.245025 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rrn75"] Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.281973 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.286805 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-94c8d"] Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.288016 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.305424 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-94c8d"] Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.388885 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/833ab86e-9608-4c6d-a546-ce8190c8fde2-config\") pod \"dnsmasq-dns-57d769cc4f-94c8d\" (UID: \"833ab86e-9608-4c6d-a546-ce8190c8fde2\") " pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.388963 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/833ab86e-9608-4c6d-a546-ce8190c8fde2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-94c8d\" (UID: \"833ab86e-9608-4c6d-a546-ce8190c8fde2\") " pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.388990 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72fvx\" (UniqueName: \"kubernetes.io/projected/833ab86e-9608-4c6d-a546-ce8190c8fde2-kube-api-access-72fvx\") pod \"dnsmasq-dns-57d769cc4f-94c8d\" (UID: \"833ab86e-9608-4c6d-a546-ce8190c8fde2\") " pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.416397 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.416445 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.490600 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/833ab86e-9608-4c6d-a546-ce8190c8fde2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-94c8d\" (UID: \"833ab86e-9608-4c6d-a546-ce8190c8fde2\") " pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.490872 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72fvx\" (UniqueName: \"kubernetes.io/projected/833ab86e-9608-4c6d-a546-ce8190c8fde2-kube-api-access-72fvx\") pod \"dnsmasq-dns-57d769cc4f-94c8d\" (UID: \"833ab86e-9608-4c6d-a546-ce8190c8fde2\") " pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.490933 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/833ab86e-9608-4c6d-a546-ce8190c8fde2-config\") pod \"dnsmasq-dns-57d769cc4f-94c8d\" (UID: \"833ab86e-9608-4c6d-a546-ce8190c8fde2\") " pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.491491 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/833ab86e-9608-4c6d-a546-ce8190c8fde2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-94c8d\" (UID: \"833ab86e-9608-4c6d-a546-ce8190c8fde2\") " pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.491783 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/833ab86e-9608-4c6d-a546-ce8190c8fde2-config\") pod \"dnsmasq-dns-57d769cc4f-94c8d\" (UID: \"833ab86e-9608-4c6d-a546-ce8190c8fde2\") " pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.526042 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72fvx\" (UniqueName: \"kubernetes.io/projected/833ab86e-9608-4c6d-a546-ce8190c8fde2-kube-api-access-72fvx\") pod \"dnsmasq-dns-57d769cc4f-94c8d\" (UID: \"833ab86e-9608-4c6d-a546-ce8190c8fde2\") " pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.682166 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" Oct 01 16:14:23 crc kubenswrapper[4726]: I1001 16:14:23.879687 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qsbql"] Oct 01 16:14:23 crc kubenswrapper[4726]: W1001 16:14:23.889613 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47538946_adc6_42ea_a62a_3cf2fb6a3206.slice/crio-890ca581c9fae19ca9932f4aa3f240b86bf7352ad6a3d426290fb4be7540e273 WatchSource:0}: Error finding container 890ca581c9fae19ca9932f4aa3f240b86bf7352ad6a3d426290fb4be7540e273: Status 404 returned error can't find the container with id 890ca581c9fae19ca9932f4aa3f240b86bf7352ad6a3d426290fb4be7540e273 Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.121153 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-94c8d"] Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.128193 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.129727 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.134338 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.134491 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-b2sq4" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.134655 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.134943 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.135174 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 01 16:14:24 crc kubenswrapper[4726]: W1001 16:14:24.135381 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod833ab86e_9608_4c6d_a546_ce8190c8fde2.slice/crio-7d85733ec80a621d774a8f50b33698f2dde9258523e65129369f0a4fcac2dc3d WatchSource:0}: Error finding container 7d85733ec80a621d774a8f50b33698f2dde9258523e65129369f0a4fcac2dc3d: Status 404 returned error can't find the container with id 7d85733ec80a621d774a8f50b33698f2dde9258523e65129369f0a4fcac2dc3d Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.136326 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.136530 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.138377 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.273145 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qsbql" event={"ID":"47538946-adc6-42ea-a62a-3cf2fb6a3206","Type":"ContainerStarted","Data":"890ca581c9fae19ca9932f4aa3f240b86bf7352ad6a3d426290fb4be7540e273"} Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.274496 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" event={"ID":"833ab86e-9608-4c6d-a546-ce8190c8fde2","Type":"ContainerStarted","Data":"7d85733ec80a621d774a8f50b33698f2dde9258523e65129369f0a4fcac2dc3d"} Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.305919 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.306020 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.306073 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/57593043-8bb5-4967-8a6c-1a55a3e8894c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.306094 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.306116 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.306716 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.306871 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qf5k\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-kube-api-access-6qf5k\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.306980 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-config-data\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.307102 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/57593043-8bb5-4967-8a6c-1a55a3e8894c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.307144 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.307358 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.408695 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.408750 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.408770 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/57593043-8bb5-4967-8a6c-1a55a3e8894c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.408787 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.408802 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.408819 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.408838 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qf5k\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-kube-api-access-6qf5k\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.408865 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-config-data\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.408889 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.408906 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/57593043-8bb5-4967-8a6c-1a55a3e8894c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.408936 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.409235 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.409524 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.410028 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.411025 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.413926 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-config-data\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.416609 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/57593043-8bb5-4967-8a6c-1a55a3e8894c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.418182 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.418933 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.421161 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/57593043-8bb5-4967-8a6c-1a55a3e8894c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.432681 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.440434 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qf5k\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-kube-api-access-6qf5k\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.466546 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.468060 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.479418 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.479539 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.479538 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.480545 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.480754 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.480836 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.481283 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-g5k8t" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.496882 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.611884 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.611978 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k92rn\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-kube-api-access-k92rn\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.612023 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.612119 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75dbebbf-c4f3-4508-901d-05148fd48f74-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.612162 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.612200 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.612248 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75dbebbf-c4f3-4508-901d-05148fd48f74-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.612275 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.612299 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.612322 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.612376 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.614637 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.714550 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.714631 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k92rn\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-kube-api-access-k92rn\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.714686 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.714732 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75dbebbf-c4f3-4508-901d-05148fd48f74-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.714784 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.714812 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.714864 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75dbebbf-c4f3-4508-901d-05148fd48f74-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.715862 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.715861 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.715975 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.716585 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.714896 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.716682 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.716710 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.716754 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.717017 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.717477 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.722108 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75dbebbf-c4f3-4508-901d-05148fd48f74-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.723416 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75dbebbf-c4f3-4508-901d-05148fd48f74-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.724012 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.726712 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.739343 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k92rn\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-kube-api-access-k92rn\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.742888 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.788484 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 16:14:24 crc kubenswrapper[4726]: I1001 16:14:24.834274 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:14:25 crc kubenswrapper[4726]: I1001 16:14:25.288962 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:14:25 crc kubenswrapper[4726]: W1001 16:14:25.319705 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57593043_8bb5_4967_8a6c_1a55a3e8894c.slice/crio-5080ba6f47df3de6a066783f373eaa1014b63a486e183c8bf31f9c881e568b04 WatchSource:0}: Error finding container 5080ba6f47df3de6a066783f373eaa1014b63a486e183c8bf31f9c881e568b04: Status 404 returned error can't find the container with id 5080ba6f47df3de6a066783f373eaa1014b63a486e183c8bf31f9c881e568b04 Oct 01 16:14:25 crc kubenswrapper[4726]: I1001 16:14:25.397123 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:14:25 crc kubenswrapper[4726]: W1001 16:14:25.420160 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75dbebbf_c4f3_4508_901d_05148fd48f74.slice/crio-787e96333c99acae7d2829130a3c13c0c3569051acd1fa48211b55442a8b83c8 WatchSource:0}: Error finding container 787e96333c99acae7d2829130a3c13c0c3569051acd1fa48211b55442a8b83c8: Status 404 returned error can't find the container with id 787e96333c99acae7d2829130a3c13c0c3569051acd1fa48211b55442a8b83c8 Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.004581 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.006982 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.009563 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-ccdvd" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.009894 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.010359 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.015237 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.018468 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.019556 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.020373 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.138275 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d743678c-e92d-4515-88d0-27e4a5086aac-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.138331 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.138557 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d743678c-e92d-4515-88d0-27e4a5086aac-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.138639 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d743678c-e92d-4515-88d0-27e4a5086aac-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.138767 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d743678c-e92d-4515-88d0-27e4a5086aac-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.138802 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb965\" (UniqueName: \"kubernetes.io/projected/d743678c-e92d-4515-88d0-27e4a5086aac-kube-api-access-kb965\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.138851 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d743678c-e92d-4515-88d0-27e4a5086aac-config-data-default\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.138912 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d743678c-e92d-4515-88d0-27e4a5086aac-kolla-config\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.138941 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d743678c-e92d-4515-88d0-27e4a5086aac-secrets\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.239966 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d743678c-e92d-4515-88d0-27e4a5086aac-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.240018 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d743678c-e92d-4515-88d0-27e4a5086aac-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.240117 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d743678c-e92d-4515-88d0-27e4a5086aac-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.240136 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb965\" (UniqueName: \"kubernetes.io/projected/d743678c-e92d-4515-88d0-27e4a5086aac-kube-api-access-kb965\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.240164 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d743678c-e92d-4515-88d0-27e4a5086aac-config-data-default\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.240187 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d743678c-e92d-4515-88d0-27e4a5086aac-kolla-config\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.240205 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d743678c-e92d-4515-88d0-27e4a5086aac-secrets\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.240233 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d743678c-e92d-4515-88d0-27e4a5086aac-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.240251 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.240570 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.242027 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d743678c-e92d-4515-88d0-27e4a5086aac-config-data-default\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.248305 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d743678c-e92d-4515-88d0-27e4a5086aac-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.248428 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d743678c-e92d-4515-88d0-27e4a5086aac-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.249766 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d743678c-e92d-4515-88d0-27e4a5086aac-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.250427 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d743678c-e92d-4515-88d0-27e4a5086aac-kolla-config\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.250846 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d743678c-e92d-4515-88d0-27e4a5086aac-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.257524 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d743678c-e92d-4515-88d0-27e4a5086aac-secrets\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.262139 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.267806 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb965\" (UniqueName: \"kubernetes.io/projected/d743678c-e92d-4515-88d0-27e4a5086aac-kube-api-access-kb965\") pod \"openstack-galera-0\" (UID: \"d743678c-e92d-4515-88d0-27e4a5086aac\") " pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.290443 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"57593043-8bb5-4967-8a6c-1a55a3e8894c","Type":"ContainerStarted","Data":"5080ba6f47df3de6a066783f373eaa1014b63a486e183c8bf31f9c881e568b04"} Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.292032 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75dbebbf-c4f3-4508-901d-05148fd48f74","Type":"ContainerStarted","Data":"787e96333c99acae7d2829130a3c13c0c3569051acd1fa48211b55442a8b83c8"} Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.328836 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.793279 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.859818 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.861903 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.863513 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-ds6d8" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.863917 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.863902 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.864852 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.882865 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.956019 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e08387ed-ffb4-4488-9345-02504e1c66cc-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.956182 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08387ed-ffb4-4488-9345-02504e1c66cc-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.956271 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e08387ed-ffb4-4488-9345-02504e1c66cc-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.956346 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.956420 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e08387ed-ffb4-4488-9345-02504e1c66cc-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.956439 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e08387ed-ffb4-4488-9345-02504e1c66cc-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.956473 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e08387ed-ffb4-4488-9345-02504e1c66cc-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.956566 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e08387ed-ffb4-4488-9345-02504e1c66cc-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:26 crc kubenswrapper[4726]: I1001 16:14:26.956589 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krtjz\" (UniqueName: \"kubernetes.io/projected/e08387ed-ffb4-4488-9345-02504e1c66cc-kube-api-access-krtjz\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.058005 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krtjz\" (UniqueName: \"kubernetes.io/projected/e08387ed-ffb4-4488-9345-02504e1c66cc-kube-api-access-krtjz\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.058081 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e08387ed-ffb4-4488-9345-02504e1c66cc-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.058111 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08387ed-ffb4-4488-9345-02504e1c66cc-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.058132 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e08387ed-ffb4-4488-9345-02504e1c66cc-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.058161 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.058195 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e08387ed-ffb4-4488-9345-02504e1c66cc-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.058211 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e08387ed-ffb4-4488-9345-02504e1c66cc-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.058243 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e08387ed-ffb4-4488-9345-02504e1c66cc-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.058274 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e08387ed-ffb4-4488-9345-02504e1c66cc-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.059089 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e08387ed-ffb4-4488-9345-02504e1c66cc-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.059100 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e08387ed-ffb4-4488-9345-02504e1c66cc-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.059239 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.059877 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e08387ed-ffb4-4488-9345-02504e1c66cc-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.060380 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e08387ed-ffb4-4488-9345-02504e1c66cc-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.070478 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08387ed-ffb4-4488-9345-02504e1c66cc-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.073518 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e08387ed-ffb4-4488-9345-02504e1c66cc-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.074111 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e08387ed-ffb4-4488-9345-02504e1c66cc-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.079800 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krtjz\" (UniqueName: \"kubernetes.io/projected/e08387ed-ffb4-4488-9345-02504e1c66cc-kube-api-access-krtjz\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.088585 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e08387ed-ffb4-4488-9345-02504e1c66cc\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.151861 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.153083 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.157179 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.157283 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.157543 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-sp65t" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.167941 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.184583 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.267843 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd24681e-73eb-49d0-bd62-4d11a485c18e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.267890 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fd24681e-73eb-49d0-bd62-4d11a485c18e-kolla-config\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.267912 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd24681e-73eb-49d0-bd62-4d11a485c18e-config-data\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.267948 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x75h\" (UniqueName: \"kubernetes.io/projected/fd24681e-73eb-49d0-bd62-4d11a485c18e-kube-api-access-2x75h\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.267986 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd24681e-73eb-49d0-bd62-4d11a485c18e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.306727 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d743678c-e92d-4515-88d0-27e4a5086aac","Type":"ContainerStarted","Data":"d4b84770c5dda4f1c9387023b13a99f85837bf01e5493e31f74403dd7de8ecbf"} Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.370205 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd24681e-73eb-49d0-bd62-4d11a485c18e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.370325 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd24681e-73eb-49d0-bd62-4d11a485c18e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.370363 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fd24681e-73eb-49d0-bd62-4d11a485c18e-kolla-config\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.370416 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd24681e-73eb-49d0-bd62-4d11a485c18e-config-data\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.370485 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x75h\" (UniqueName: \"kubernetes.io/projected/fd24681e-73eb-49d0-bd62-4d11a485c18e-kube-api-access-2x75h\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.372506 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fd24681e-73eb-49d0-bd62-4d11a485c18e-kolla-config\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.372571 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd24681e-73eb-49d0-bd62-4d11a485c18e-config-data\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.386650 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd24681e-73eb-49d0-bd62-4d11a485c18e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.388268 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x75h\" (UniqueName: \"kubernetes.io/projected/fd24681e-73eb-49d0-bd62-4d11a485c18e-kube-api-access-2x75h\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.395619 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd24681e-73eb-49d0-bd62-4d11a485c18e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fd24681e-73eb-49d0-bd62-4d11a485c18e\") " pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.493307 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.658312 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 16:14:27 crc kubenswrapper[4726]: W1001 16:14:27.951777 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd24681e_73eb_49d0_bd62_4d11a485c18e.slice/crio-dd30a912bc72c39ede29de0aadb7d13088dd8b9f35df4ef192aec22fdc1b356c WatchSource:0}: Error finding container dd30a912bc72c39ede29de0aadb7d13088dd8b9f35df4ef192aec22fdc1b356c: Status 404 returned error can't find the container with id dd30a912bc72c39ede29de0aadb7d13088dd8b9f35df4ef192aec22fdc1b356c Oct 01 16:14:27 crc kubenswrapper[4726]: I1001 16:14:27.952396 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 01 16:14:28 crc kubenswrapper[4726]: I1001 16:14:28.318999 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e08387ed-ffb4-4488-9345-02504e1c66cc","Type":"ContainerStarted","Data":"7e176f047deb7c156ef8d1088f640c7e047749badd180ed29b224690d51e1aea"} Oct 01 16:14:28 crc kubenswrapper[4726]: I1001 16:14:28.321919 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fd24681e-73eb-49d0-bd62-4d11a485c18e","Type":"ContainerStarted","Data":"dd30a912bc72c39ede29de0aadb7d13088dd8b9f35df4ef192aec22fdc1b356c"} Oct 01 16:14:29 crc kubenswrapper[4726]: I1001 16:14:29.052462 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:14:29 crc kubenswrapper[4726]: I1001 16:14:29.053693 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 16:14:29 crc kubenswrapper[4726]: I1001 16:14:29.056556 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-nqrk2" Oct 01 16:14:29 crc kubenswrapper[4726]: I1001 16:14:29.060159 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:14:29 crc kubenswrapper[4726]: I1001 16:14:29.199040 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chp9b\" (UniqueName: \"kubernetes.io/projected/9639916d-aee7-485b-960f-c57fb48af4a0-kube-api-access-chp9b\") pod \"kube-state-metrics-0\" (UID: \"9639916d-aee7-485b-960f-c57fb48af4a0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:14:29 crc kubenswrapper[4726]: I1001 16:14:29.301093 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chp9b\" (UniqueName: \"kubernetes.io/projected/9639916d-aee7-485b-960f-c57fb48af4a0-kube-api-access-chp9b\") pod \"kube-state-metrics-0\" (UID: \"9639916d-aee7-485b-960f-c57fb48af4a0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:14:29 crc kubenswrapper[4726]: I1001 16:14:29.319673 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chp9b\" (UniqueName: \"kubernetes.io/projected/9639916d-aee7-485b-960f-c57fb48af4a0-kube-api-access-chp9b\") pod \"kube-state-metrics-0\" (UID: \"9639916d-aee7-485b-960f-c57fb48af4a0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:14:29 crc kubenswrapper[4726]: I1001 16:14:29.379212 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 16:14:32 crc kubenswrapper[4726]: E1001 16:14:32.486174 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = reading blob sha256:0b0b5d8b2d4a7273a90d53aa5a624050895f1a6e86e71cde8da95b67be91a595: Digest did not match, expected sha256:0b0b5d8b2d4a7273a90d53aa5a624050895f1a6e86e71cde8da95b67be91a595, got sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Oct 01 16:14:32 crc kubenswrapper[4726]: E1001 16:14:32.486625 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k92rn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(75dbebbf-c4f3-4508-901d-05148fd48f74): ErrImagePull: reading blob sha256:0b0b5d8b2d4a7273a90d53aa5a624050895f1a6e86e71cde8da95b67be91a595: Digest did not match, expected sha256:0b0b5d8b2d4a7273a90d53aa5a624050895f1a6e86e71cde8da95b67be91a595, got sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" logger="UnhandledError" Oct 01 16:14:32 crc kubenswrapper[4726]: E1001 16:14:32.487809 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"reading blob sha256:0b0b5d8b2d4a7273a90d53aa5a624050895f1a6e86e71cde8da95b67be91a595: Digest did not match, expected sha256:0b0b5d8b2d4a7273a90d53aa5a624050895f1a6e86e71cde8da95b67be91a595, got sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855\"" pod="openstack/rabbitmq-cell1-server-0" podUID="75dbebbf-c4f3-4508-901d-05148fd48f74" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.502989 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-z4lk4"] Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.503937 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.505769 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-rnbhr" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.506078 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.506338 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.518700 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z4lk4"] Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.575261 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-4msxz"] Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.576902 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.580115 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-4msxz"] Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.648335 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzm9b\" (UniqueName: \"kubernetes.io/projected/525c9172-edd7-4483-b649-128a2d7ac283-kube-api-access-rzm9b\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.648426 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f67ec6c1-adf7-47f9-855f-fde38e0430c7-var-run-ovn\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.648476 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f67ec6c1-adf7-47f9-855f-fde38e0430c7-var-run\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.648499 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f67ec6c1-adf7-47f9-855f-fde38e0430c7-scripts\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.648530 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f67ec6c1-adf7-47f9-855f-fde38e0430c7-ovn-controller-tls-certs\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.648656 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk7z4\" (UniqueName: \"kubernetes.io/projected/f67ec6c1-adf7-47f9-855f-fde38e0430c7-kube-api-access-lk7z4\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.648839 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/525c9172-edd7-4483-b649-128a2d7ac283-scripts\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.648963 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f67ec6c1-adf7-47f9-855f-fde38e0430c7-combined-ca-bundle\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.649132 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/525c9172-edd7-4483-b649-128a2d7ac283-var-run\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.649202 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/525c9172-edd7-4483-b649-128a2d7ac283-var-lib\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.649229 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/525c9172-edd7-4483-b649-128a2d7ac283-var-log\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.649284 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f67ec6c1-adf7-47f9-855f-fde38e0430c7-var-log-ovn\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.649339 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/525c9172-edd7-4483-b649-128a2d7ac283-etc-ovs\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.754428 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzm9b\" (UniqueName: \"kubernetes.io/projected/525c9172-edd7-4483-b649-128a2d7ac283-kube-api-access-rzm9b\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.754483 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f67ec6c1-adf7-47f9-855f-fde38e0430c7-var-run-ovn\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.754510 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f67ec6c1-adf7-47f9-855f-fde38e0430c7-var-run\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.754527 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f67ec6c1-adf7-47f9-855f-fde38e0430c7-scripts\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.754550 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f67ec6c1-adf7-47f9-855f-fde38e0430c7-ovn-controller-tls-certs\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.754578 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk7z4\" (UniqueName: \"kubernetes.io/projected/f67ec6c1-adf7-47f9-855f-fde38e0430c7-kube-api-access-lk7z4\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.754595 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/525c9172-edd7-4483-b649-128a2d7ac283-scripts\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.754624 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f67ec6c1-adf7-47f9-855f-fde38e0430c7-combined-ca-bundle\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.754661 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/525c9172-edd7-4483-b649-128a2d7ac283-var-run\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.754689 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/525c9172-edd7-4483-b649-128a2d7ac283-var-lib\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.754706 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f67ec6c1-adf7-47f9-855f-fde38e0430c7-var-log-ovn\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.754724 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/525c9172-edd7-4483-b649-128a2d7ac283-var-log\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.754743 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/525c9172-edd7-4483-b649-128a2d7ac283-etc-ovs\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.755117 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/525c9172-edd7-4483-b649-128a2d7ac283-etc-ovs\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.755206 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f67ec6c1-adf7-47f9-855f-fde38e0430c7-var-run\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.755199 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/525c9172-edd7-4483-b649-128a2d7ac283-var-run\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.755309 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/525c9172-edd7-4483-b649-128a2d7ac283-var-lib\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.755327 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f67ec6c1-adf7-47f9-855f-fde38e0430c7-var-run-ovn\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.755405 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f67ec6c1-adf7-47f9-855f-fde38e0430c7-var-log-ovn\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.755490 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/525c9172-edd7-4483-b649-128a2d7ac283-var-log\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.758210 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/525c9172-edd7-4483-b649-128a2d7ac283-scripts\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.758769 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f67ec6c1-adf7-47f9-855f-fde38e0430c7-scripts\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.766106 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f67ec6c1-adf7-47f9-855f-fde38e0430c7-ovn-controller-tls-certs\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.766112 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f67ec6c1-adf7-47f9-855f-fde38e0430c7-combined-ca-bundle\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.770937 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzm9b\" (UniqueName: \"kubernetes.io/projected/525c9172-edd7-4483-b649-128a2d7ac283-kube-api-access-rzm9b\") pod \"ovn-controller-ovs-4msxz\" (UID: \"525c9172-edd7-4483-b649-128a2d7ac283\") " pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.776042 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk7z4\" (UniqueName: \"kubernetes.io/projected/f67ec6c1-adf7-47f9-855f-fde38e0430c7-kube-api-access-lk7z4\") pod \"ovn-controller-z4lk4\" (UID: \"f67ec6c1-adf7-47f9-855f-fde38e0430c7\") " pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.827815 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z4lk4" Oct 01 16:14:32 crc kubenswrapper[4726]: I1001 16:14:32.896510 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.115220 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.116916 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.119554 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-vl55n" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.119820 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.120064 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.120748 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.120881 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.124259 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.261819 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.261874 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.261902 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.261917 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl85j\" (UniqueName: \"kubernetes.io/projected/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-kube-api-access-wl85j\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.261941 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.261969 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.262026 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-config\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.262066 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: E1001 16:14:33.360208 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="75dbebbf-c4f3-4508-901d-05148fd48f74" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.363436 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-config\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.363479 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.363521 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.363542 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.363564 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.363579 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl85j\" (UniqueName: \"kubernetes.io/projected/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-kube-api-access-wl85j\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.363601 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.363626 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.364041 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.366151 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-config\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.366577 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.367183 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.369395 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.370108 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.388477 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl85j\" (UniqueName: \"kubernetes.io/projected/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-kube-api-access-wl85j\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.390264 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.407278 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:33 crc kubenswrapper[4726]: I1001 16:14:33.437502 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.860973 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.863445 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.867168 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.867481 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.867664 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-bxt52" Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.868467 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.883588 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.924243 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/23600bcf-22d0-4de1-be32-5adf413748b4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.924574 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23600bcf-22d0-4de1-be32-5adf413748b4-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.924893 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/23600bcf-22d0-4de1-be32-5adf413748b4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.925000 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xz8j\" (UniqueName: \"kubernetes.io/projected/23600bcf-22d0-4de1-be32-5adf413748b4-kube-api-access-9xz8j\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.925117 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.925216 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23600bcf-22d0-4de1-be32-5adf413748b4-config\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.925351 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/23600bcf-22d0-4de1-be32-5adf413748b4-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:36 crc kubenswrapper[4726]: I1001 16:14:36.925430 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23600bcf-22d0-4de1-be32-5adf413748b4-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.026944 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/23600bcf-22d0-4de1-be32-5adf413748b4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.027073 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23600bcf-22d0-4de1-be32-5adf413748b4-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.027098 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/23600bcf-22d0-4de1-be32-5adf413748b4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.027124 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xz8j\" (UniqueName: \"kubernetes.io/projected/23600bcf-22d0-4de1-be32-5adf413748b4-kube-api-access-9xz8j\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.027194 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.027261 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23600bcf-22d0-4de1-be32-5adf413748b4-config\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.027310 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/23600bcf-22d0-4de1-be32-5adf413748b4-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.027339 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23600bcf-22d0-4de1-be32-5adf413748b4-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.027726 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.028556 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/23600bcf-22d0-4de1-be32-5adf413748b4-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.028888 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23600bcf-22d0-4de1-be32-5adf413748b4-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.029391 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23600bcf-22d0-4de1-be32-5adf413748b4-config\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.052991 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.651490 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/23600bcf-22d0-4de1-be32-5adf413748b4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.651492 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/23600bcf-22d0-4de1-be32-5adf413748b4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.651584 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23600bcf-22d0-4de1-be32-5adf413748b4-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.656955 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xz8j\" (UniqueName: \"kubernetes.io/projected/23600bcf-22d0-4de1-be32-5adf413748b4-kube-api-access-9xz8j\") pod \"ovsdbserver-sb-0\" (UID: \"23600bcf-22d0-4de1-be32-5adf413748b4\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:37 crc kubenswrapper[4726]: I1001 16:14:37.808044 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 01 16:14:42 crc kubenswrapper[4726]: E1001 16:14:42.751318 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 01 16:14:42 crc kubenswrapper[4726]: E1001 16:14:42.752081 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vv4jw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-rrn75_openstack(0d7affd8-5a89-4ec3-a593-b4dfe592861f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:14:42 crc kubenswrapper[4726]: E1001 16:14:42.753479 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" podUID="0d7affd8-5a89-4ec3-a593-b4dfe592861f" Oct 01 16:14:42 crc kubenswrapper[4726]: E1001 16:14:42.846262 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 01 16:14:42 crc kubenswrapper[4726]: E1001 16:14:42.846452 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tls67,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-mxntn_openstack(add42128-3a49-40ae-baec-e74f857203fc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:14:42 crc kubenswrapper[4726]: E1001 16:14:42.847652 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-mxntn" podUID="add42128-3a49-40ae-baec-e74f857203fc" Oct 01 16:14:43 crc kubenswrapper[4726]: E1001 16:14:43.555247 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 01 16:14:43 crc kubenswrapper[4726]: E1001 16:14:43.556495 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-72fvx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-94c8d_openstack(833ab86e-9608-4c6d-a546-ce8190c8fde2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:14:43 crc kubenswrapper[4726]: E1001 16:14:43.558188 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" podUID="833ab86e-9608-4c6d-a546-ce8190c8fde2" Oct 01 16:14:43 crc kubenswrapper[4726]: I1001 16:14:43.994158 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:14:44 crc kubenswrapper[4726]: I1001 16:14:44.144521 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jnxwz"] Oct 01 16:14:44 crc kubenswrapper[4726]: I1001 16:14:44.146279 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:14:44 crc kubenswrapper[4726]: I1001 16:14:44.157346 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jnxwz"] Oct 01 16:14:44 crc kubenswrapper[4726]: I1001 16:14:44.253359 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12562596-fc27-4e0f-84b1-f06ee267cbe5-utilities\") pod \"community-operators-jnxwz\" (UID: \"12562596-fc27-4e0f-84b1-f06ee267cbe5\") " pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:14:44 crc kubenswrapper[4726]: I1001 16:14:44.253451 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12562596-fc27-4e0f-84b1-f06ee267cbe5-catalog-content\") pod \"community-operators-jnxwz\" (UID: \"12562596-fc27-4e0f-84b1-f06ee267cbe5\") " pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:14:44 crc kubenswrapper[4726]: I1001 16:14:44.253504 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hlgm\" (UniqueName: \"kubernetes.io/projected/12562596-fc27-4e0f-84b1-f06ee267cbe5-kube-api-access-4hlgm\") pod \"community-operators-jnxwz\" (UID: \"12562596-fc27-4e0f-84b1-f06ee267cbe5\") " pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:14:44 crc kubenswrapper[4726]: I1001 16:14:44.354976 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12562596-fc27-4e0f-84b1-f06ee267cbe5-utilities\") pod \"community-operators-jnxwz\" (UID: \"12562596-fc27-4e0f-84b1-f06ee267cbe5\") " pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:14:44 crc kubenswrapper[4726]: I1001 16:14:44.355080 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12562596-fc27-4e0f-84b1-f06ee267cbe5-catalog-content\") pod \"community-operators-jnxwz\" (UID: \"12562596-fc27-4e0f-84b1-f06ee267cbe5\") " pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:14:44 crc kubenswrapper[4726]: I1001 16:14:44.355128 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hlgm\" (UniqueName: \"kubernetes.io/projected/12562596-fc27-4e0f-84b1-f06ee267cbe5-kube-api-access-4hlgm\") pod \"community-operators-jnxwz\" (UID: \"12562596-fc27-4e0f-84b1-f06ee267cbe5\") " pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:14:44 crc kubenswrapper[4726]: I1001 16:14:44.356190 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12562596-fc27-4e0f-84b1-f06ee267cbe5-utilities\") pod \"community-operators-jnxwz\" (UID: \"12562596-fc27-4e0f-84b1-f06ee267cbe5\") " pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:14:44 crc kubenswrapper[4726]: I1001 16:14:44.356211 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12562596-fc27-4e0f-84b1-f06ee267cbe5-catalog-content\") pod \"community-operators-jnxwz\" (UID: \"12562596-fc27-4e0f-84b1-f06ee267cbe5\") " pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:14:44 crc kubenswrapper[4726]: I1001 16:14:44.389190 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hlgm\" (UniqueName: \"kubernetes.io/projected/12562596-fc27-4e0f-84b1-f06ee267cbe5-kube-api-access-4hlgm\") pod \"community-operators-jnxwz\" (UID: \"12562596-fc27-4e0f-84b1-f06ee267cbe5\") " pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:14:44 crc kubenswrapper[4726]: E1001 16:14:44.464309 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" podUID="833ab86e-9608-4c6d-a546-ce8190c8fde2" Oct 01 16:14:44 crc kubenswrapper[4726]: I1001 16:14:44.487713 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:14:46 crc kubenswrapper[4726]: W1001 16:14:46.621541 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9639916d_aee7_485b_960f_c57fb48af4a0.slice/crio-bab572d58892c3670a8e18157e982a146e27b07bbb4d25f66ae72e9834e67959 WatchSource:0}: Error finding container bab572d58892c3670a8e18157e982a146e27b07bbb4d25f66ae72e9834e67959: Status 404 returned error can't find the container with id bab572d58892c3670a8e18157e982a146e27b07bbb4d25f66ae72e9834e67959 Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.678811 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mxntn" Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.684250 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.685373 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.791814 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vv4jw\" (UniqueName: \"kubernetes.io/projected/0d7affd8-5a89-4ec3-a593-b4dfe592861f-kube-api-access-vv4jw\") pod \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\" (UID: \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\") " Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.791856 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d7affd8-5a89-4ec3-a593-b4dfe592861f-config\") pod \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\" (UID: \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\") " Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.791891 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tls67\" (UniqueName: \"kubernetes.io/projected/add42128-3a49-40ae-baec-e74f857203fc-kube-api-access-tls67\") pod \"add42128-3a49-40ae-baec-e74f857203fc\" (UID: \"add42128-3a49-40ae-baec-e74f857203fc\") " Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.791911 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d7affd8-5a89-4ec3-a593-b4dfe592861f-dns-svc\") pod \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\" (UID: \"0d7affd8-5a89-4ec3-a593-b4dfe592861f\") " Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.792002 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add42128-3a49-40ae-baec-e74f857203fc-config\") pod \"add42128-3a49-40ae-baec-e74f857203fc\" (UID: \"add42128-3a49-40ae-baec-e74f857203fc\") " Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.792777 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/add42128-3a49-40ae-baec-e74f857203fc-config" (OuterVolumeSpecName: "config") pod "add42128-3a49-40ae-baec-e74f857203fc" (UID: "add42128-3a49-40ae-baec-e74f857203fc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.792773 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d7affd8-5a89-4ec3-a593-b4dfe592861f-config" (OuterVolumeSpecName: "config") pod "0d7affd8-5a89-4ec3-a593-b4dfe592861f" (UID: "0d7affd8-5a89-4ec3-a593-b4dfe592861f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.793106 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d7affd8-5a89-4ec3-a593-b4dfe592861f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0d7affd8-5a89-4ec3-a593-b4dfe592861f" (UID: "0d7affd8-5a89-4ec3-a593-b4dfe592861f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.796655 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d7affd8-5a89-4ec3-a593-b4dfe592861f-kube-api-access-vv4jw" (OuterVolumeSpecName: "kube-api-access-vv4jw") pod "0d7affd8-5a89-4ec3-a593-b4dfe592861f" (UID: "0d7affd8-5a89-4ec3-a593-b4dfe592861f"). InnerVolumeSpecName "kube-api-access-vv4jw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.796901 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/add42128-3a49-40ae-baec-e74f857203fc-kube-api-access-tls67" (OuterVolumeSpecName: "kube-api-access-tls67") pod "add42128-3a49-40ae-baec-e74f857203fc" (UID: "add42128-3a49-40ae-baec-e74f857203fc"). InnerVolumeSpecName "kube-api-access-tls67". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.893943 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d7affd8-5a89-4ec3-a593-b4dfe592861f-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.894575 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tls67\" (UniqueName: \"kubernetes.io/projected/add42128-3a49-40ae-baec-e74f857203fc-kube-api-access-tls67\") on node \"crc\" DevicePath \"\"" Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.894731 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d7affd8-5a89-4ec3-a593-b4dfe592861f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.895041 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add42128-3a49-40ae-baec-e74f857203fc-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:14:46 crc kubenswrapper[4726]: I1001 16:14:46.895259 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vv4jw\" (UniqueName: \"kubernetes.io/projected/0d7affd8-5a89-4ec3-a593-b4dfe592861f-kube-api-access-vv4jw\") on node \"crc\" DevicePath \"\"" Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.406775 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z4lk4"] Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.452614 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 16:14:47 crc kubenswrapper[4726]: W1001 16:14:47.461191 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23600bcf_22d0_4de1_be32_5adf413748b4.slice/crio-5494cdca4dbca9c93ff2c2245e5eee7a6452feded80c8fe370d9dc3359dfdaaf WatchSource:0}: Error finding container 5494cdca4dbca9c93ff2c2245e5eee7a6452feded80c8fe370d9dc3359dfdaaf: Status 404 returned error can't find the container with id 5494cdca4dbca9c93ff2c2245e5eee7a6452feded80c8fe370d9dc3359dfdaaf Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.481948 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d743678c-e92d-4515-88d0-27e4a5086aac","Type":"ContainerStarted","Data":"23560e986175d0630bcaab925431cc752c4dbd2097324fad76e47be3765b11b3"} Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.483857 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e08387ed-ffb4-4488-9345-02504e1c66cc","Type":"ContainerStarted","Data":"7c54a5c3bff4e24e86f43def64b19e940b7c0879b22a17cfe4675867cd2f26d6"} Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.485143 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" event={"ID":"0d7affd8-5a89-4ec3-a593-b4dfe592861f","Type":"ContainerDied","Data":"acac541f6a01907ab4cb6040e09ec89c22435e64d40428e61af76478e10ef95b"} Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.485222 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rrn75" Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.486515 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"23600bcf-22d0-4de1-be32-5adf413748b4","Type":"ContainerStarted","Data":"5494cdca4dbca9c93ff2c2245e5eee7a6452feded80c8fe370d9dc3359dfdaaf"} Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.487861 4726 generic.go:334] "Generic (PLEG): container finished" podID="47538946-adc6-42ea-a62a-3cf2fb6a3206" containerID="4c701161a5ed766c6ce3aff83c08c52cad5ac6acf4bbf04ff0440cb72f53a473" exitCode=0 Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.487916 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qsbql" event={"ID":"47538946-adc6-42ea-a62a-3cf2fb6a3206","Type":"ContainerDied","Data":"4c701161a5ed766c6ce3aff83c08c52cad5ac6acf4bbf04ff0440cb72f53a473"} Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.489492 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9639916d-aee7-485b-960f-c57fb48af4a0","Type":"ContainerStarted","Data":"bab572d58892c3670a8e18157e982a146e27b07bbb4d25f66ae72e9834e67959"} Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.490509 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-mxntn" event={"ID":"add42128-3a49-40ae-baec-e74f857203fc","Type":"ContainerDied","Data":"58b75bccfa69b77129cdf6900ad4cd2f2a432bacfeb4b8b2397d8b43e861508e"} Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.490530 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mxntn" Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.493232 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fd24681e-73eb-49d0-bd62-4d11a485c18e","Type":"ContainerStarted","Data":"97fa6b7ba7609acf3f2ead9ab43cd4aaae5414209b505e5bfdd826d24c274e6a"} Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.493312 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.494358 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z4lk4" event={"ID":"f67ec6c1-adf7-47f9-855f-fde38e0430c7","Type":"ContainerStarted","Data":"bec28d24d469560ecc9b7d287ee7a3eb2579e2042bd9b73278ac1ef5eca4495c"} Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.546816 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jnxwz"] Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.558136 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=1.635315122 podStartE2EDuration="20.558111531s" podCreationTimestamp="2025-10-01 16:14:27 +0000 UTC" firstStartedPulling="2025-10-01 16:14:27.957728965 +0000 UTC m=+920.859281542" lastFinishedPulling="2025-10-01 16:14:46.880525374 +0000 UTC m=+939.782077951" observedRunningTime="2025-10-01 16:14:47.55290448 +0000 UTC m=+940.454457077" watchObservedRunningTime="2025-10-01 16:14:47.558111531 +0000 UTC m=+940.459664108" Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.584897 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 16:14:47 crc kubenswrapper[4726]: W1001 16:14:47.590568 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b88b062_70c9_4ea5_9df5_d7eafcc3c6ab.slice/crio-12ca49690842ef4d496911fc463d53d01ca167576ba9b708842e6f7b116f7154 WatchSource:0}: Error finding container 12ca49690842ef4d496911fc463d53d01ca167576ba9b708842e6f7b116f7154: Status 404 returned error can't find the container with id 12ca49690842ef4d496911fc463d53d01ca167576ba9b708842e6f7b116f7154 Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.622846 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rrn75"] Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.652118 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rrn75"] Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.665164 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mxntn"] Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.670695 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mxntn"] Oct 01 16:14:47 crc kubenswrapper[4726]: E1001 16:14:47.733769 4726 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Oct 01 16:14:47 crc kubenswrapper[4726]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/47538946-adc6-42ea-a62a-3cf2fb6a3206/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 01 16:14:47 crc kubenswrapper[4726]: > podSandboxID="890ca581c9fae19ca9932f4aa3f240b86bf7352ad6a3d426290fb4be7540e273" Oct 01 16:14:47 crc kubenswrapper[4726]: E1001 16:14:47.733941 4726 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 01 16:14:47 crc kubenswrapper[4726]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5spq9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-qsbql_openstack(47538946-adc6-42ea-a62a-3cf2fb6a3206): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/47538946-adc6-42ea-a62a-3cf2fb6a3206/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 01 16:14:47 crc kubenswrapper[4726]: > logger="UnhandledError" Oct 01 16:14:47 crc kubenswrapper[4726]: E1001 16:14:47.735178 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/47538946-adc6-42ea-a62a-3cf2fb6a3206/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-666b6646f7-qsbql" podUID="47538946-adc6-42ea-a62a-3cf2fb6a3206" Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.820298 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d7affd8-5a89-4ec3-a593-b4dfe592861f" path="/var/lib/kubelet/pods/0d7affd8-5a89-4ec3-a593-b4dfe592861f/volumes" Oct 01 16:14:47 crc kubenswrapper[4726]: I1001 16:14:47.820730 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="add42128-3a49-40ae-baec-e74f857203fc" path="/var/lib/kubelet/pods/add42128-3a49-40ae-baec-e74f857203fc/volumes" Oct 01 16:14:48 crc kubenswrapper[4726]: I1001 16:14:48.506010 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"57593043-8bb5-4967-8a6c-1a55a3e8894c","Type":"ContainerStarted","Data":"2deafad37316839f98e372793ca49f200759055adcb7e9e7e730bfd5d7b19193"} Oct 01 16:14:48 crc kubenswrapper[4726]: I1001 16:14:48.508228 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab","Type":"ContainerStarted","Data":"12ca49690842ef4d496911fc463d53d01ca167576ba9b708842e6f7b116f7154"} Oct 01 16:14:48 crc kubenswrapper[4726]: I1001 16:14:48.510681 4726 generic.go:334] "Generic (PLEG): container finished" podID="12562596-fc27-4e0f-84b1-f06ee267cbe5" containerID="0c19bbd94879716e829aa57d6a696869d19a0232c791904e60388fcd0ef129cc" exitCode=0 Oct 01 16:14:48 crc kubenswrapper[4726]: I1001 16:14:48.512286 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnxwz" event={"ID":"12562596-fc27-4e0f-84b1-f06ee267cbe5","Type":"ContainerDied","Data":"0c19bbd94879716e829aa57d6a696869d19a0232c791904e60388fcd0ef129cc"} Oct 01 16:14:48 crc kubenswrapper[4726]: I1001 16:14:48.512441 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnxwz" event={"ID":"12562596-fc27-4e0f-84b1-f06ee267cbe5","Type":"ContainerStarted","Data":"23f9158931427d4720b9f6e2a175452294ecf324f57b98858a48051a2bde3f73"} Oct 01 16:14:48 crc kubenswrapper[4726]: I1001 16:14:48.600099 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-4msxz"] Oct 01 16:14:48 crc kubenswrapper[4726]: W1001 16:14:48.994629 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod525c9172_edd7_4483_b649_128a2d7ac283.slice/crio-d1cb58ef1f45cd8cbe0221b8609518fb1b276f385379f8cd32ade1d058fe4606 WatchSource:0}: Error finding container d1cb58ef1f45cd8cbe0221b8609518fb1b276f385379f8cd32ade1d058fe4606: Status 404 returned error can't find the container with id d1cb58ef1f45cd8cbe0221b8609518fb1b276f385379f8cd32ade1d058fe4606 Oct 01 16:14:49 crc kubenswrapper[4726]: I1001 16:14:49.523635 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75dbebbf-c4f3-4508-901d-05148fd48f74","Type":"ContainerStarted","Data":"198fb4a2291752198257d14ae7da6349daa4258c3f409a6a3f85c2d745acc28d"} Oct 01 16:14:49 crc kubenswrapper[4726]: I1001 16:14:49.526911 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qsbql" event={"ID":"47538946-adc6-42ea-a62a-3cf2fb6a3206","Type":"ContainerStarted","Data":"39fcd9cffb0b4be1dabd72b8d8c128df909bc9caeff126760a573dda836b0bfc"} Oct 01 16:14:49 crc kubenswrapper[4726]: I1001 16:14:49.527912 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:14:49 crc kubenswrapper[4726]: I1001 16:14:49.542784 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4msxz" event={"ID":"525c9172-edd7-4483-b649-128a2d7ac283","Type":"ContainerStarted","Data":"d1cb58ef1f45cd8cbe0221b8609518fb1b276f385379f8cd32ade1d058fe4606"} Oct 01 16:14:49 crc kubenswrapper[4726]: I1001 16:14:49.577373 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-qsbql" podStartSLOduration=4.577173343 podStartE2EDuration="27.577357842s" podCreationTimestamp="2025-10-01 16:14:22 +0000 UTC" firstStartedPulling="2025-10-01 16:14:23.891939722 +0000 UTC m=+916.793492299" lastFinishedPulling="2025-10-01 16:14:46.892124221 +0000 UTC m=+939.793676798" observedRunningTime="2025-10-01 16:14:49.57452685 +0000 UTC m=+942.476079427" watchObservedRunningTime="2025-10-01 16:14:49.577357842 +0000 UTC m=+942.478910419" Oct 01 16:14:52 crc kubenswrapper[4726]: I1001 16:14:52.494185 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 01 16:14:53 crc kubenswrapper[4726]: I1001 16:14:53.413442 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:14:53 crc kubenswrapper[4726]: I1001 16:14:53.413522 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:14:53 crc kubenswrapper[4726]: I1001 16:14:53.582541 4726 generic.go:334] "Generic (PLEG): container finished" podID="d743678c-e92d-4515-88d0-27e4a5086aac" containerID="23560e986175d0630bcaab925431cc752c4dbd2097324fad76e47be3765b11b3" exitCode=0 Oct 01 16:14:53 crc kubenswrapper[4726]: I1001 16:14:53.582624 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d743678c-e92d-4515-88d0-27e4a5086aac","Type":"ContainerDied","Data":"23560e986175d0630bcaab925431cc752c4dbd2097324fad76e47be3765b11b3"} Oct 01 16:14:53 crc kubenswrapper[4726]: I1001 16:14:53.584546 4726 generic.go:334] "Generic (PLEG): container finished" podID="e08387ed-ffb4-4488-9345-02504e1c66cc" containerID="7c54a5c3bff4e24e86f43def64b19e940b7c0879b22a17cfe4675867cd2f26d6" exitCode=0 Oct 01 16:14:53 crc kubenswrapper[4726]: I1001 16:14:53.584586 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e08387ed-ffb4-4488-9345-02504e1c66cc","Type":"ContainerDied","Data":"7c54a5c3bff4e24e86f43def64b19e940b7c0879b22a17cfe4675867cd2f26d6"} Oct 01 16:14:58 crc kubenswrapper[4726]: I1001 16:14:58.283216 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.418782 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-94c8d"] Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.505136 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-g4lff"] Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.510609 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.521544 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-g4lff"] Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.637464 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e08387ed-ffb4-4488-9345-02504e1c66cc","Type":"ContainerStarted","Data":"140ca6a2db0fb221d32c7dc8eab8af040506f3bc1136f024040484c9c33e8c19"} Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.637672 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-config\") pod \"dnsmasq-dns-7cb5889db5-g4lff\" (UID: \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\") " pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.637722 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qnsc\" (UniqueName: \"kubernetes.io/projected/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-kube-api-access-4qnsc\") pod \"dnsmasq-dns-7cb5889db5-g4lff\" (UID: \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\") " pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.637771 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-g4lff\" (UID: \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\") " pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.665407 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=15.263051908 podStartE2EDuration="34.665389974s" podCreationTimestamp="2025-10-01 16:14:25 +0000 UTC" firstStartedPulling="2025-10-01 16:14:27.673539252 +0000 UTC m=+920.575091829" lastFinishedPulling="2025-10-01 16:14:47.075877318 +0000 UTC m=+939.977429895" observedRunningTime="2025-10-01 16:14:59.663969913 +0000 UTC m=+952.565522480" watchObservedRunningTime="2025-10-01 16:14:59.665389974 +0000 UTC m=+952.566942551" Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.738702 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-config\") pod \"dnsmasq-dns-7cb5889db5-g4lff\" (UID: \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\") " pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.738766 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qnsc\" (UniqueName: \"kubernetes.io/projected/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-kube-api-access-4qnsc\") pod \"dnsmasq-dns-7cb5889db5-g4lff\" (UID: \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\") " pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.738844 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-g4lff\" (UID: \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\") " pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.740230 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-g4lff\" (UID: \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\") " pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.740996 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-config\") pod \"dnsmasq-dns-7cb5889db5-g4lff\" (UID: \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\") " pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" Oct 01 16:14:59 crc kubenswrapper[4726]: I1001 16:14:59.767993 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qnsc\" (UniqueName: \"kubernetes.io/projected/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-kube-api-access-4qnsc\") pod \"dnsmasq-dns-7cb5889db5-g4lff\" (UID: \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\") " pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.043964 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.132137 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q"] Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.133603 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.139722 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.144099 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q"] Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.146605 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.175684 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b99038f-6158-4768-ad4e-b6e9cec7bf72-config-volume\") pod \"collect-profiles-29322255-tgl6q\" (UID: \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.176022 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b99038f-6158-4768-ad4e-b6e9cec7bf72-secret-volume\") pod \"collect-profiles-29322255-tgl6q\" (UID: \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.277601 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b99038f-6158-4768-ad4e-b6e9cec7bf72-config-volume\") pod \"collect-profiles-29322255-tgl6q\" (UID: \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.277673 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b99038f-6158-4768-ad4e-b6e9cec7bf72-secret-volume\") pod \"collect-profiles-29322255-tgl6q\" (UID: \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.277783 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8f8g\" (UniqueName: \"kubernetes.io/projected/0b99038f-6158-4768-ad4e-b6e9cec7bf72-kube-api-access-b8f8g\") pod \"collect-profiles-29322255-tgl6q\" (UID: \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.278495 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b99038f-6158-4768-ad4e-b6e9cec7bf72-config-volume\") pod \"collect-profiles-29322255-tgl6q\" (UID: \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.284674 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b99038f-6158-4768-ad4e-b6e9cec7bf72-secret-volume\") pod \"collect-profiles-29322255-tgl6q\" (UID: \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.379221 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8f8g\" (UniqueName: \"kubernetes.io/projected/0b99038f-6158-4768-ad4e-b6e9cec7bf72-kube-api-access-b8f8g\") pod \"collect-profiles-29322255-tgl6q\" (UID: \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.401640 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8f8g\" (UniqueName: \"kubernetes.io/projected/0b99038f-6158-4768-ad4e-b6e9cec7bf72-kube-api-access-b8f8g\") pod \"collect-profiles-29322255-tgl6q\" (UID: \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.470657 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.523887 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-g4lff"] Oct 01 16:15:00 crc kubenswrapper[4726]: W1001 16:15:00.534771 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28a68ce6_bbf6_4e46_8414_05ff3ed1a23f.slice/crio-bea3a868171b6db28c225360e932a99e38725d6859f8814b072b20b18b4ae5a6 WatchSource:0}: Error finding container bea3a868171b6db28c225360e932a99e38725d6859f8814b072b20b18b4ae5a6: Status 404 returned error can't find the container with id bea3a868171b6db28c225360e932a99e38725d6859f8814b072b20b18b4ae5a6 Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.556382 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.565968 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.568865 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-69vmk" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.569488 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.569491 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.569861 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.576920 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.656176 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" event={"ID":"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f","Type":"ContainerStarted","Data":"bea3a868171b6db28c225360e932a99e38725d6859f8814b072b20b18b4ae5a6"} Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.657908 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab","Type":"ContainerStarted","Data":"3a374fb602ad3924399faae5f3158570d47e642878b4a6477e308a1ec4e97516"} Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.686433 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.686479 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.686498 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6d192df1-7269-4493-b17b-4037addc226e-lock\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.686516 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sswzv\" (UniqueName: \"kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-kube-api-access-sswzv\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.686620 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6d192df1-7269-4493-b17b-4037addc226e-cache\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.787534 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6d192df1-7269-4493-b17b-4037addc226e-cache\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.787589 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.787613 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.787636 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6d192df1-7269-4493-b17b-4037addc226e-lock\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.787652 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sswzv\" (UniqueName: \"kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-kube-api-access-sswzv\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.787977 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.788682 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6d192df1-7269-4493-b17b-4037addc226e-lock\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.788740 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6d192df1-7269-4493-b17b-4037addc226e-cache\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: E1001 16:15:00.796193 4726 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 16:15:00 crc kubenswrapper[4726]: E1001 16:15:00.796249 4726 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 16:15:00 crc kubenswrapper[4726]: E1001 16:15:00.796304 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift podName:6d192df1-7269-4493-b17b-4037addc226e nodeName:}" failed. No retries permitted until 2025-10-01 16:15:01.296287716 +0000 UTC m=+954.197840293 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift") pod "swift-storage-0" (UID: "6d192df1-7269-4493-b17b-4037addc226e") : configmap "swift-ring-files" not found Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.808488 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sswzv\" (UniqueName: \"kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-kube-api-access-sswzv\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.812957 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:00 crc kubenswrapper[4726]: I1001 16:15:00.952622 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q"] Oct 01 16:15:00 crc kubenswrapper[4726]: W1001 16:15:00.960386 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b99038f_6158_4768_ad4e_b6e9cec7bf72.slice/crio-f5a2f64ef5d970c3d42907d776d5071ecdde9d35533ff7160af2f84b232f7414 WatchSource:0}: Error finding container f5a2f64ef5d970c3d42907d776d5071ecdde9d35533ff7160af2f84b232f7414: Status 404 returned error can't find the container with id f5a2f64ef5d970c3d42907d776d5071ecdde9d35533ff7160af2f84b232f7414 Oct 01 16:15:01 crc kubenswrapper[4726]: I1001 16:15:01.298468 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:01 crc kubenswrapper[4726]: E1001 16:15:01.298645 4726 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 16:15:01 crc kubenswrapper[4726]: E1001 16:15:01.298662 4726 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 16:15:01 crc kubenswrapper[4726]: E1001 16:15:01.298721 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift podName:6d192df1-7269-4493-b17b-4037addc226e nodeName:}" failed. No retries permitted until 2025-10-01 16:15:02.298699037 +0000 UTC m=+955.200251634 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift") pod "swift-storage-0" (UID: "6d192df1-7269-4493-b17b-4037addc226e") : configmap "swift-ring-files" not found Oct 01 16:15:01 crc kubenswrapper[4726]: I1001 16:15:01.668000 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" event={"ID":"833ab86e-9608-4c6d-a546-ce8190c8fde2","Type":"ContainerStarted","Data":"6c327322e48ac29c4c34a6477a238fd83de4392f06685371d0e1f68a82530eac"} Oct 01 16:15:01 crc kubenswrapper[4726]: I1001 16:15:01.669863 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"23600bcf-22d0-4de1-be32-5adf413748b4","Type":"ContainerStarted","Data":"c1561270dd2853670b195b6d17a1c90ac42237e6d13e55470df3049978f5e875"} Oct 01 16:15:01 crc kubenswrapper[4726]: I1001 16:15:01.672134 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4msxz" event={"ID":"525c9172-edd7-4483-b649-128a2d7ac283","Type":"ContainerStarted","Data":"5ae5df19f6f0721c79524cd4e5229f531c96408fb1abb986c1ee20d6b6d7cd5d"} Oct 01 16:15:01 crc kubenswrapper[4726]: I1001 16:15:01.673915 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" event={"ID":"0b99038f-6158-4768-ad4e-b6e9cec7bf72","Type":"ContainerStarted","Data":"f5a2f64ef5d970c3d42907d776d5071ecdde9d35533ff7160af2f84b232f7414"} Oct 01 16:15:01 crc kubenswrapper[4726]: I1001 16:15:01.676103 4726 generic.go:334] "Generic (PLEG): container finished" podID="12562596-fc27-4e0f-84b1-f06ee267cbe5" containerID="db4e3de37114da59675196f0388b9b45602f37d30d92e8683923aed099b5e05f" exitCode=0 Oct 01 16:15:01 crc kubenswrapper[4726]: I1001 16:15:01.676128 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnxwz" event={"ID":"12562596-fc27-4e0f-84b1-f06ee267cbe5","Type":"ContainerDied","Data":"db4e3de37114da59675196f0388b9b45602f37d30d92e8683923aed099b5e05f"} Oct 01 16:15:01 crc kubenswrapper[4726]: I1001 16:15:01.678142 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d743678c-e92d-4515-88d0-27e4a5086aac","Type":"ContainerStarted","Data":"c9d60ab74d537363e6498b5039447934cf30b70a153d81a7d303f508f4ff6572"} Oct 01 16:15:01 crc kubenswrapper[4726]: I1001 16:15:01.679676 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z4lk4" event={"ID":"f67ec6c1-adf7-47f9-855f-fde38e0430c7","Type":"ContainerStarted","Data":"272ef097f7490269ed8f4fb8c4a3d7d70357532267ad5d6028d12b37242c4a3a"} Oct 01 16:15:02 crc kubenswrapper[4726]: E1001 16:15:02.309533 4726 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.217:42584->38.102.83.217:36913: write tcp 38.102.83.217:42584->38.102.83.217:36913: write: broken pipe Oct 01 16:15:02 crc kubenswrapper[4726]: I1001 16:15:02.316148 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:02 crc kubenswrapper[4726]: E1001 16:15:02.316333 4726 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 16:15:02 crc kubenswrapper[4726]: E1001 16:15:02.316354 4726 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 16:15:02 crc kubenswrapper[4726]: E1001 16:15:02.316391 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift podName:6d192df1-7269-4493-b17b-4037addc226e nodeName:}" failed. No retries permitted until 2025-10-01 16:15:04.31637693 +0000 UTC m=+957.217929507 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift") pod "swift-storage-0" (UID: "6d192df1-7269-4493-b17b-4037addc226e") : configmap "swift-ring-files" not found Oct 01 16:15:02 crc kubenswrapper[4726]: I1001 16:15:02.689131 4726 generic.go:334] "Generic (PLEG): container finished" podID="833ab86e-9608-4c6d-a546-ce8190c8fde2" containerID="6c327322e48ac29c4c34a6477a238fd83de4392f06685371d0e1f68a82530eac" exitCode=0 Oct 01 16:15:02 crc kubenswrapper[4726]: I1001 16:15:02.689217 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" event={"ID":"833ab86e-9608-4c6d-a546-ce8190c8fde2","Type":"ContainerDied","Data":"6c327322e48ac29c4c34a6477a238fd83de4392f06685371d0e1f68a82530eac"} Oct 01 16:15:02 crc kubenswrapper[4726]: I1001 16:15:02.691574 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" event={"ID":"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f","Type":"ContainerStarted","Data":"e7c8114eb0341cbd293f2d63b089410bdc6432b67ad17ced326d0337f00ad3c1"} Oct 01 16:15:02 crc kubenswrapper[4726]: I1001 16:15:02.693292 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" event={"ID":"0b99038f-6158-4768-ad4e-b6e9cec7bf72","Type":"ContainerStarted","Data":"17d9a341b2fae3ddffbbbd1b780de16a97302f2fb7fea899d568340653c1bfa1"} Oct 01 16:15:02 crc kubenswrapper[4726]: I1001 16:15:02.693504 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-z4lk4" Oct 01 16:15:02 crc kubenswrapper[4726]: I1001 16:15:02.734586 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-z4lk4" podStartSLOduration=20.508048649 podStartE2EDuration="30.734563984s" podCreationTimestamp="2025-10-01 16:14:32 +0000 UTC" firstStartedPulling="2025-10-01 16:14:47.415967973 +0000 UTC m=+940.317520540" lastFinishedPulling="2025-10-01 16:14:57.642483298 +0000 UTC m=+950.544035875" observedRunningTime="2025-10-01 16:15:02.734329098 +0000 UTC m=+955.635881685" watchObservedRunningTime="2025-10-01 16:15:02.734563984 +0000 UTC m=+955.636116561" Oct 01 16:15:03 crc kubenswrapper[4726]: I1001 16:15:03.727514 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" podStartSLOduration=3.72749633 podStartE2EDuration="3.72749633s" podCreationTimestamp="2025-10-01 16:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:15:03.724537924 +0000 UTC m=+956.626090541" watchObservedRunningTime="2025-10-01 16:15:03.72749633 +0000 UTC m=+956.629048907" Oct 01 16:15:03 crc kubenswrapper[4726]: I1001 16:15:03.776632 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=19.719270098 podStartE2EDuration="39.776611096s" podCreationTimestamp="2025-10-01 16:14:24 +0000 UTC" firstStartedPulling="2025-10-01 16:14:26.835143444 +0000 UTC m=+919.736696021" lastFinishedPulling="2025-10-01 16:14:46.892484422 +0000 UTC m=+939.794037019" observedRunningTime="2025-10-01 16:15:03.769002545 +0000 UTC m=+956.670555132" watchObservedRunningTime="2025-10-01 16:15:03.776611096 +0000 UTC m=+956.678163683" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.036208 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.146386 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/833ab86e-9608-4c6d-a546-ce8190c8fde2-config\") pod \"833ab86e-9608-4c6d-a546-ce8190c8fde2\" (UID: \"833ab86e-9608-4c6d-a546-ce8190c8fde2\") " Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.146451 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/833ab86e-9608-4c6d-a546-ce8190c8fde2-dns-svc\") pod \"833ab86e-9608-4c6d-a546-ce8190c8fde2\" (UID: \"833ab86e-9608-4c6d-a546-ce8190c8fde2\") " Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.146541 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72fvx\" (UniqueName: \"kubernetes.io/projected/833ab86e-9608-4c6d-a546-ce8190c8fde2-kube-api-access-72fvx\") pod \"833ab86e-9608-4c6d-a546-ce8190c8fde2\" (UID: \"833ab86e-9608-4c6d-a546-ce8190c8fde2\") " Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.171878 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/833ab86e-9608-4c6d-a546-ce8190c8fde2-kube-api-access-72fvx" (OuterVolumeSpecName: "kube-api-access-72fvx") pod "833ab86e-9608-4c6d-a546-ce8190c8fde2" (UID: "833ab86e-9608-4c6d-a546-ce8190c8fde2"). InnerVolumeSpecName "kube-api-access-72fvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.175095 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/833ab86e-9608-4c6d-a546-ce8190c8fde2-config" (OuterVolumeSpecName: "config") pod "833ab86e-9608-4c6d-a546-ce8190c8fde2" (UID: "833ab86e-9608-4c6d-a546-ce8190c8fde2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.191472 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/833ab86e-9608-4c6d-a546-ce8190c8fde2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "833ab86e-9608-4c6d-a546-ce8190c8fde2" (UID: "833ab86e-9608-4c6d-a546-ce8190c8fde2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.248418 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/833ab86e-9608-4c6d-a546-ce8190c8fde2-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.248463 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/833ab86e-9608-4c6d-a546-ce8190c8fde2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.248527 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72fvx\" (UniqueName: \"kubernetes.io/projected/833ab86e-9608-4c6d-a546-ce8190c8fde2-kube-api-access-72fvx\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.350181 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:04 crc kubenswrapper[4726]: E1001 16:15:04.350410 4726 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 16:15:04 crc kubenswrapper[4726]: E1001 16:15:04.350445 4726 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 16:15:04 crc kubenswrapper[4726]: E1001 16:15:04.350524 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift podName:6d192df1-7269-4493-b17b-4037addc226e nodeName:}" failed. No retries permitted until 2025-10-01 16:15:08.350499843 +0000 UTC m=+961.252052440 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift") pod "swift-storage-0" (UID: "6d192df1-7269-4493-b17b-4037addc226e") : configmap "swift-ring-files" not found Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.666461 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-t4hqx"] Oct 01 16:15:04 crc kubenswrapper[4726]: E1001 16:15:04.666996 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="833ab86e-9608-4c6d-a546-ce8190c8fde2" containerName="init" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.667030 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="833ab86e-9608-4c6d-a546-ce8190c8fde2" containerName="init" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.667345 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="833ab86e-9608-4c6d-a546-ce8190c8fde2" containerName="init" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.668207 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.670801 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.671081 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.671220 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.675596 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-t4hqx"] Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.712496 4726 generic.go:334] "Generic (PLEG): container finished" podID="525c9172-edd7-4483-b649-128a2d7ac283" containerID="5ae5df19f6f0721c79524cd4e5229f531c96408fb1abb986c1ee20d6b6d7cd5d" exitCode=0 Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.712605 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4msxz" event={"ID":"525c9172-edd7-4483-b649-128a2d7ac283","Type":"ContainerDied","Data":"5ae5df19f6f0721c79524cd4e5229f531c96408fb1abb986c1ee20d6b6d7cd5d"} Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.716715 4726 generic.go:334] "Generic (PLEG): container finished" podID="28a68ce6-bbf6-4e46-8414-05ff3ed1a23f" containerID="e7c8114eb0341cbd293f2d63b089410bdc6432b67ad17ced326d0337f00ad3c1" exitCode=0 Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.716799 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" event={"ID":"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f","Type":"ContainerDied","Data":"e7c8114eb0341cbd293f2d63b089410bdc6432b67ad17ced326d0337f00ad3c1"} Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.721768 4726 generic.go:334] "Generic (PLEG): container finished" podID="0b99038f-6158-4768-ad4e-b6e9cec7bf72" containerID="17d9a341b2fae3ddffbbbd1b780de16a97302f2fb7fea899d568340653c1bfa1" exitCode=0 Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.721924 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" event={"ID":"0b99038f-6158-4768-ad4e-b6e9cec7bf72","Type":"ContainerDied","Data":"17d9a341b2fae3ddffbbbd1b780de16a97302f2fb7fea899d568340653c1bfa1"} Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.729111 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" event={"ID":"833ab86e-9608-4c6d-a546-ce8190c8fde2","Type":"ContainerDied","Data":"7d85733ec80a621d774a8f50b33698f2dde9258523e65129369f0a4fcac2dc3d"} Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.729175 4726 scope.go:117] "RemoveContainer" containerID="6c327322e48ac29c4c34a6477a238fd83de4392f06685371d0e1f68a82530eac" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.729351 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-94c8d" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.757973 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4ab0d371-af59-407e-b596-3c533ad031b8-ring-data-devices\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.758020 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4ab0d371-af59-407e-b596-3c533ad031b8-etc-swift\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.758080 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-combined-ca-bundle\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.758105 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ab0d371-af59-407e-b596-3c533ad031b8-scripts\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.758136 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5ld6\" (UniqueName: \"kubernetes.io/projected/4ab0d371-af59-407e-b596-3c533ad031b8-kube-api-access-x5ld6\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.758281 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-dispersionconf\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.758517 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-swiftconf\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.821400 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-94c8d"] Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.826983 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-94c8d"] Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.859917 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5ld6\" (UniqueName: \"kubernetes.io/projected/4ab0d371-af59-407e-b596-3c533ad031b8-kube-api-access-x5ld6\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.861471 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-dispersionconf\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.868241 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-swiftconf\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.868429 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4ab0d371-af59-407e-b596-3c533ad031b8-ring-data-devices\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.868505 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4ab0d371-af59-407e-b596-3c533ad031b8-etc-swift\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.868603 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-combined-ca-bundle\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.868663 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ab0d371-af59-407e-b596-3c533ad031b8-scripts\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.870376 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4ab0d371-af59-407e-b596-3c533ad031b8-ring-data-devices\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.870884 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4ab0d371-af59-407e-b596-3c533ad031b8-etc-swift\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.872766 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ab0d371-af59-407e-b596-3c533ad031b8-scripts\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.874247 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-swiftconf\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.874315 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-dispersionconf\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.876524 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-combined-ca-bundle\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.890588 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5ld6\" (UniqueName: \"kubernetes.io/projected/4ab0d371-af59-407e-b596-3c533ad031b8-kube-api-access-x5ld6\") pod \"swift-ring-rebalance-t4hqx\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:04 crc kubenswrapper[4726]: I1001 16:15:04.998357 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:05 crc kubenswrapper[4726]: I1001 16:15:05.476116 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-t4hqx"] Oct 01 16:15:05 crc kubenswrapper[4726]: I1001 16:15:05.821182 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="833ab86e-9608-4c6d-a546-ce8190c8fde2" path="/var/lib/kubelet/pods/833ab86e-9608-4c6d-a546-ce8190c8fde2/volumes" Oct 01 16:15:06 crc kubenswrapper[4726]: I1001 16:15:06.328979 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 01 16:15:06 crc kubenswrapper[4726]: I1001 16:15:06.329294 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 01 16:15:06 crc kubenswrapper[4726]: I1001 16:15:06.754760 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4msxz" event={"ID":"525c9172-edd7-4483-b649-128a2d7ac283","Type":"ContainerStarted","Data":"f1870aabc2f5da3b172a54755918b23a25f2e00ad4685c361dd642c34c03892b"} Oct 01 16:15:07 crc kubenswrapper[4726]: I1001 16:15:07.185668 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 01 16:15:07 crc kubenswrapper[4726]: I1001 16:15:07.185758 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 01 16:15:08 crc kubenswrapper[4726]: I1001 16:15:08.443442 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:08 crc kubenswrapper[4726]: E1001 16:15:08.443915 4726 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 16:15:08 crc kubenswrapper[4726]: E1001 16:15:08.443940 4726 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 16:15:08 crc kubenswrapper[4726]: E1001 16:15:08.444002 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift podName:6d192df1-7269-4493-b17b-4037addc226e nodeName:}" failed. No retries permitted until 2025-10-01 16:15:16.44398052 +0000 UTC m=+969.345533097 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift") pod "swift-storage-0" (UID: "6d192df1-7269-4493-b17b-4037addc226e") : configmap "swift-ring-files" not found Oct 01 16:15:13 crc kubenswrapper[4726]: I1001 16:15:13.706179 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" Oct 01 16:15:13 crc kubenswrapper[4726]: I1001 16:15:13.821480 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t4hqx" event={"ID":"4ab0d371-af59-407e-b596-3c533ad031b8","Type":"ContainerStarted","Data":"be9528e08f651009d87981c904062ede63770d83b92a4f9e3b3175e6a861df3b"} Oct 01 16:15:13 crc kubenswrapper[4726]: I1001 16:15:13.821919 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" event={"ID":"0b99038f-6158-4768-ad4e-b6e9cec7bf72","Type":"ContainerDied","Data":"f5a2f64ef5d970c3d42907d776d5071ecdde9d35533ff7160af2f84b232f7414"} Oct 01 16:15:13 crc kubenswrapper[4726]: I1001 16:15:13.821960 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5a2f64ef5d970c3d42907d776d5071ecdde9d35533ff7160af2f84b232f7414" Oct 01 16:15:13 crc kubenswrapper[4726]: I1001 16:15:13.821968 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q" Oct 01 16:15:13 crc kubenswrapper[4726]: I1001 16:15:13.854769 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8f8g\" (UniqueName: \"kubernetes.io/projected/0b99038f-6158-4768-ad4e-b6e9cec7bf72-kube-api-access-b8f8g\") pod \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\" (UID: \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\") " Oct 01 16:15:13 crc kubenswrapper[4726]: I1001 16:15:13.854882 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b99038f-6158-4768-ad4e-b6e9cec7bf72-secret-volume\") pod \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\" (UID: \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\") " Oct 01 16:15:13 crc kubenswrapper[4726]: I1001 16:15:13.854957 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b99038f-6158-4768-ad4e-b6e9cec7bf72-config-volume\") pod \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\" (UID: \"0b99038f-6158-4768-ad4e-b6e9cec7bf72\") " Oct 01 16:15:13 crc kubenswrapper[4726]: I1001 16:15:13.856850 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b99038f-6158-4768-ad4e-b6e9cec7bf72-config-volume" (OuterVolumeSpecName: "config-volume") pod "0b99038f-6158-4768-ad4e-b6e9cec7bf72" (UID: "0b99038f-6158-4768-ad4e-b6e9cec7bf72"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:13 crc kubenswrapper[4726]: I1001 16:15:13.861666 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b99038f-6158-4768-ad4e-b6e9cec7bf72-kube-api-access-b8f8g" (OuterVolumeSpecName: "kube-api-access-b8f8g") pod "0b99038f-6158-4768-ad4e-b6e9cec7bf72" (UID: "0b99038f-6158-4768-ad4e-b6e9cec7bf72"). InnerVolumeSpecName "kube-api-access-b8f8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:13 crc kubenswrapper[4726]: I1001 16:15:13.862180 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b99038f-6158-4768-ad4e-b6e9cec7bf72-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0b99038f-6158-4768-ad4e-b6e9cec7bf72" (UID: "0b99038f-6158-4768-ad4e-b6e9cec7bf72"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:15:13 crc kubenswrapper[4726]: I1001 16:15:13.956748 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8f8g\" (UniqueName: \"kubernetes.io/projected/0b99038f-6158-4768-ad4e-b6e9cec7bf72-kube-api-access-b8f8g\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:13 crc kubenswrapper[4726]: I1001 16:15:13.956779 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b99038f-6158-4768-ad4e-b6e9cec7bf72-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:13 crc kubenswrapper[4726]: I1001 16:15:13.956788 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b99038f-6158-4768-ad4e-b6e9cec7bf72-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:16 crc kubenswrapper[4726]: I1001 16:15:16.500648 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:16 crc kubenswrapper[4726]: E1001 16:15:16.501671 4726 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 16:15:16 crc kubenswrapper[4726]: E1001 16:15:16.501703 4726 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 16:15:16 crc kubenswrapper[4726]: E1001 16:15:16.501766 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift podName:6d192df1-7269-4493-b17b-4037addc226e nodeName:}" failed. No retries permitted until 2025-10-01 16:15:32.50174611 +0000 UTC m=+985.403298697 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift") pod "swift-storage-0" (UID: "6d192df1-7269-4493-b17b-4037addc226e") : configmap "swift-ring-files" not found Oct 01 16:15:18 crc kubenswrapper[4726]: I1001 16:15:18.930428 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-h598x"] Oct 01 16:15:18 crc kubenswrapper[4726]: E1001 16:15:18.930941 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b99038f-6158-4768-ad4e-b6e9cec7bf72" containerName="collect-profiles" Oct 01 16:15:18 crc kubenswrapper[4726]: I1001 16:15:18.930952 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b99038f-6158-4768-ad4e-b6e9cec7bf72" containerName="collect-profiles" Oct 01 16:15:18 crc kubenswrapper[4726]: I1001 16:15:18.931114 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b99038f-6158-4768-ad4e-b6e9cec7bf72" containerName="collect-profiles" Oct 01 16:15:18 crc kubenswrapper[4726]: I1001 16:15:18.931616 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:18 crc kubenswrapper[4726]: I1001 16:15:18.941308 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 01 16:15:18 crc kubenswrapper[4726]: I1001 16:15:18.949853 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-h598x"] Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.051764 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/72fa255e-484d-43ab-a837-d31c20e8e4d4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.051883 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72fa255e-484d-43ab-a837-d31c20e8e4d4-combined-ca-bundle\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.052010 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72fa255e-484d-43ab-a837-d31c20e8e4d4-config\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.052110 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/72fa255e-484d-43ab-a837-d31c20e8e4d4-ovn-rundir\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.052213 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/72fa255e-484d-43ab-a837-d31c20e8e4d4-ovs-rundir\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.052316 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfkc2\" (UniqueName: \"kubernetes.io/projected/72fa255e-484d-43ab-a837-d31c20e8e4d4-kube-api-access-bfkc2\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.099293 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-g4lff"] Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.131120 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-llwt9"] Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.132457 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.136284 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.153885 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-llwt9"] Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.154340 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72fa255e-484d-43ab-a837-d31c20e8e4d4-config\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.154500 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/72fa255e-484d-43ab-a837-d31c20e8e4d4-ovn-rundir\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.154924 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/72fa255e-484d-43ab-a837-d31c20e8e4d4-ovs-rundir\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.155097 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfkc2\" (UniqueName: \"kubernetes.io/projected/72fa255e-484d-43ab-a837-d31c20e8e4d4-kube-api-access-bfkc2\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.155253 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/72fa255e-484d-43ab-a837-d31c20e8e4d4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.155346 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72fa255e-484d-43ab-a837-d31c20e8e4d4-combined-ca-bundle\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.154874 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/72fa255e-484d-43ab-a837-d31c20e8e4d4-ovn-rundir\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.155255 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/72fa255e-484d-43ab-a837-d31c20e8e4d4-ovs-rundir\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.156992 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72fa255e-484d-43ab-a837-d31c20e8e4d4-config\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.162416 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/72fa255e-484d-43ab-a837-d31c20e8e4d4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.167251 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72fa255e-484d-43ab-a837-d31c20e8e4d4-combined-ca-bundle\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.182806 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfkc2\" (UniqueName: \"kubernetes.io/projected/72fa255e-484d-43ab-a837-d31c20e8e4d4-kube-api-access-bfkc2\") pod \"ovn-controller-metrics-h598x\" (UID: \"72fa255e-484d-43ab-a837-d31c20e8e4d4\") " pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.256820 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8cg7\" (UniqueName: \"kubernetes.io/projected/cf2e368f-8ab0-49e2-965a-bd69b48e1482-kube-api-access-c8cg7\") pod \"dnsmasq-dns-57d65f699f-llwt9\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.256869 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-llwt9\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.256921 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-dns-svc\") pod \"dnsmasq-dns-57d65f699f-llwt9\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.256951 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-config\") pod \"dnsmasq-dns-57d65f699f-llwt9\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.267589 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-h598x" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.290264 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-llwt9"] Oct 01 16:15:19 crc kubenswrapper[4726]: E1001 16:15:19.291225 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-c8cg7 ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-57d65f699f-llwt9" podUID="cf2e368f-8ab0-49e2-965a-bd69b48e1482" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.329849 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-4rwkl"] Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.334176 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.337781 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.343749 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-4rwkl"] Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.362366 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8cg7\" (UniqueName: \"kubernetes.io/projected/cf2e368f-8ab0-49e2-965a-bd69b48e1482-kube-api-access-c8cg7\") pod \"dnsmasq-dns-57d65f699f-llwt9\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.362410 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-llwt9\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.362455 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-dns-svc\") pod \"dnsmasq-dns-57d65f699f-llwt9\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.362480 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-config\") pod \"dnsmasq-dns-57d65f699f-llwt9\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.363325 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-config\") pod \"dnsmasq-dns-57d65f699f-llwt9\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.364018 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-llwt9\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.364550 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-dns-svc\") pod \"dnsmasq-dns-57d65f699f-llwt9\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.388480 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8cg7\" (UniqueName: \"kubernetes.io/projected/cf2e368f-8ab0-49e2-965a-bd69b48e1482-kube-api-access-c8cg7\") pod \"dnsmasq-dns-57d65f699f-llwt9\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.465331 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-config\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.465379 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.465414 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.465442 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.465733 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ztmm\" (UniqueName: \"kubernetes.io/projected/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-kube-api-access-6ztmm\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.566857 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.566906 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.567002 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ztmm\" (UniqueName: \"kubernetes.io/projected/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-kube-api-access-6ztmm\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.567036 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-config\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.567065 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.567804 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.567900 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-config\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.568002 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.568634 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.582086 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ztmm\" (UniqueName: \"kubernetes.io/projected/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-kube-api-access-6ztmm\") pod \"dnsmasq-dns-b8fbc5445-4rwkl\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.649413 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.771592 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-h598x"] Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.863712 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-h598x" event={"ID":"72fa255e-484d-43ab-a837-d31c20e8e4d4","Type":"ContainerStarted","Data":"d1a705ddc32a9533e7aa83474a203d2161967eebe9e36aa9b2cf3851e5199335"} Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.863729 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.874310 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.972658 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8cg7\" (UniqueName: \"kubernetes.io/projected/cf2e368f-8ab0-49e2-965a-bd69b48e1482-kube-api-access-c8cg7\") pod \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.972825 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-config\") pod \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.972858 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-dns-svc\") pod \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.972917 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-ovsdbserver-nb\") pod \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\" (UID: \"cf2e368f-8ab0-49e2-965a-bd69b48e1482\") " Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.973449 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-config" (OuterVolumeSpecName: "config") pod "cf2e368f-8ab0-49e2-965a-bd69b48e1482" (UID: "cf2e368f-8ab0-49e2-965a-bd69b48e1482"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.973556 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cf2e368f-8ab0-49e2-965a-bd69b48e1482" (UID: "cf2e368f-8ab0-49e2-965a-bd69b48e1482"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.973553 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cf2e368f-8ab0-49e2-965a-bd69b48e1482" (UID: "cf2e368f-8ab0-49e2-965a-bd69b48e1482"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:19 crc kubenswrapper[4726]: I1001 16:15:19.980340 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf2e368f-8ab0-49e2-965a-bd69b48e1482-kube-api-access-c8cg7" (OuterVolumeSpecName: "kube-api-access-c8cg7") pod "cf2e368f-8ab0-49e2-965a-bd69b48e1482" (UID: "cf2e368f-8ab0-49e2-965a-bd69b48e1482"). InnerVolumeSpecName "kube-api-access-c8cg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:20 crc kubenswrapper[4726]: I1001 16:15:20.075552 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:20 crc kubenswrapper[4726]: I1001 16:15:20.075832 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:20 crc kubenswrapper[4726]: I1001 16:15:20.075854 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf2e368f-8ab0-49e2-965a-bd69b48e1482-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:20 crc kubenswrapper[4726]: I1001 16:15:20.075886 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8cg7\" (UniqueName: \"kubernetes.io/projected/cf2e368f-8ab0-49e2-965a-bd69b48e1482-kube-api-access-c8cg7\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:20 crc kubenswrapper[4726]: I1001 16:15:20.104875 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-4rwkl"] Oct 01 16:15:20 crc kubenswrapper[4726]: W1001 16:15:20.112311 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode34367e7_1d09_41ef_ad55_e11ce0e5f2a3.slice/crio-be518624d44f984cce7f96bd4920ea8189151cbf9eed04ac82221e209a28e67f WatchSource:0}: Error finding container be518624d44f984cce7f96bd4920ea8189151cbf9eed04ac82221e209a28e67f: Status 404 returned error can't find the container with id be518624d44f984cce7f96bd4920ea8189151cbf9eed04ac82221e209a28e67f Oct 01 16:15:20 crc kubenswrapper[4726]: I1001 16:15:20.872385 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9639916d-aee7-485b-960f-c57fb48af4a0","Type":"ContainerStarted","Data":"b060292e5702871769507f89bf0f98f54b08a2b66040f7f1accfb5d82ed0d82a"} Oct 01 16:15:20 crc kubenswrapper[4726]: I1001 16:15:20.873440 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" event={"ID":"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3","Type":"ContainerStarted","Data":"be518624d44f984cce7f96bd4920ea8189151cbf9eed04ac82221e209a28e67f"} Oct 01 16:15:20 crc kubenswrapper[4726]: I1001 16:15:20.875811 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" event={"ID":"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f","Type":"ContainerStarted","Data":"cf92f6831f92d4e18a0b204c24859bc9026341ad868303d25abe34fb87869188"} Oct 01 16:15:20 crc kubenswrapper[4726]: I1001 16:15:20.877481 4726 generic.go:334] "Generic (PLEG): container finished" podID="57593043-8bb5-4967-8a6c-1a55a3e8894c" containerID="2deafad37316839f98e372793ca49f200759055adcb7e9e7e730bfd5d7b19193" exitCode=0 Oct 01 16:15:20 crc kubenswrapper[4726]: I1001 16:15:20.877606 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-llwt9" Oct 01 16:15:20 crc kubenswrapper[4726]: I1001 16:15:20.878305 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"57593043-8bb5-4967-8a6c-1a55a3e8894c","Type":"ContainerDied","Data":"2deafad37316839f98e372793ca49f200759055adcb7e9e7e730bfd5d7b19193"} Oct 01 16:15:21 crc kubenswrapper[4726]: I1001 16:15:21.004939 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-llwt9"] Oct 01 16:15:21 crc kubenswrapper[4726]: I1001 16:15:21.004996 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-llwt9"] Oct 01 16:15:21 crc kubenswrapper[4726]: I1001 16:15:21.818281 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf2e368f-8ab0-49e2-965a-bd69b48e1482" path="/var/lib/kubelet/pods/cf2e368f-8ab0-49e2-965a-bd69b48e1482/volumes" Oct 01 16:15:21 crc kubenswrapper[4726]: I1001 16:15:21.888250 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4msxz" event={"ID":"525c9172-edd7-4483-b649-128a2d7ac283","Type":"ContainerStarted","Data":"a06abaa73467ef11123501d0d13ba523cff4e2c142c227761b5726628ef440b5"} Oct 01 16:15:21 crc kubenswrapper[4726]: I1001 16:15:21.893850 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnxwz" event={"ID":"12562596-fc27-4e0f-84b1-f06ee267cbe5","Type":"ContainerStarted","Data":"7354a0ba99e8490acd9a196a901a72214e653dc109e54f6e71266f2b2cda33f6"} Oct 01 16:15:21 crc kubenswrapper[4726]: I1001 16:15:21.895763 4726 generic.go:334] "Generic (PLEG): container finished" podID="75dbebbf-c4f3-4508-901d-05148fd48f74" containerID="198fb4a2291752198257d14ae7da6349daa4258c3f409a6a3f85c2d745acc28d" exitCode=0 Oct 01 16:15:21 crc kubenswrapper[4726]: I1001 16:15:21.895804 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75dbebbf-c4f3-4508-901d-05148fd48f74","Type":"ContainerDied","Data":"198fb4a2291752198257d14ae7da6349daa4258c3f409a6a3f85c2d745acc28d"} Oct 01 16:15:21 crc kubenswrapper[4726]: I1001 16:15:21.897363 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" event={"ID":"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3","Type":"ContainerStarted","Data":"3b75a4eec0509e57bec3226a4fcd9a24e96dfba115a8e09a91fb952d564df1c2"} Oct 01 16:15:22 crc kubenswrapper[4726]: I1001 16:15:22.906787 4726 generic.go:334] "Generic (PLEG): container finished" podID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" containerID="3b75a4eec0509e57bec3226a4fcd9a24e96dfba115a8e09a91fb952d564df1c2" exitCode=0 Oct 01 16:15:22 crc kubenswrapper[4726]: I1001 16:15:22.906842 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" event={"ID":"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3","Type":"ContainerDied","Data":"3b75a4eec0509e57bec3226a4fcd9a24e96dfba115a8e09a91fb952d564df1c2"} Oct 01 16:15:22 crc kubenswrapper[4726]: I1001 16:15:22.909081 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" podUID="28a68ce6-bbf6-4e46-8414-05ff3ed1a23f" containerName="dnsmasq-dns" containerID="cri-o://cf92f6831f92d4e18a0b204c24859bc9026341ad868303d25abe34fb87869188" gracePeriod=10 Oct 01 16:15:22 crc kubenswrapper[4726]: I1001 16:15:22.909176 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"57593043-8bb5-4967-8a6c-1a55a3e8894c","Type":"ContainerStarted","Data":"8217cab599a9ea7e8ecc95c31f47ebced3aad20e93476a47c5964c4cb700cbf8"} Oct 01 16:15:22 crc kubenswrapper[4726]: I1001 16:15:22.909363 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" Oct 01 16:15:22 crc kubenswrapper[4726]: I1001 16:15:22.909409 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:15:22 crc kubenswrapper[4726]: I1001 16:15:22.909436 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:15:22 crc kubenswrapper[4726]: I1001 16:15:22.909576 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 01 16:15:22 crc kubenswrapper[4726]: I1001 16:15:22.975435 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-4msxz" podStartSLOduration=42.021638696 podStartE2EDuration="50.975415679s" podCreationTimestamp="2025-10-01 16:14:32 +0000 UTC" firstStartedPulling="2025-10-01 16:14:48.997052459 +0000 UTC m=+941.898605036" lastFinishedPulling="2025-10-01 16:14:57.950829442 +0000 UTC m=+950.852382019" observedRunningTime="2025-10-01 16:15:22.963776631 +0000 UTC m=+975.865329208" watchObservedRunningTime="2025-10-01 16:15:22.975415679 +0000 UTC m=+975.876968256" Oct 01 16:15:22 crc kubenswrapper[4726]: I1001 16:15:22.994026 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" podStartSLOduration=23.994009929 podStartE2EDuration="23.994009929s" podCreationTimestamp="2025-10-01 16:14:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:15:22.992383272 +0000 UTC m=+975.893935859" watchObservedRunningTime="2025-10-01 16:15:22.994009929 +0000 UTC m=+975.895562506" Oct 01 16:15:23 crc kubenswrapper[4726]: I1001 16:15:23.018923 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jnxwz" podStartSLOduration=8.815065737 podStartE2EDuration="39.018905752s" podCreationTimestamp="2025-10-01 16:14:44 +0000 UTC" firstStartedPulling="2025-10-01 16:14:48.769911843 +0000 UTC m=+941.671464440" lastFinishedPulling="2025-10-01 16:15:18.973751878 +0000 UTC m=+971.875304455" observedRunningTime="2025-10-01 16:15:23.017009487 +0000 UTC m=+975.918562064" watchObservedRunningTime="2025-10-01 16:15:23.018905752 +0000 UTC m=+975.920458329" Oct 01 16:15:23 crc kubenswrapper[4726]: I1001 16:15:23.032485 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=21.73289122 podStartE2EDuration="54.032469836s" podCreationTimestamp="2025-10-01 16:14:29 +0000 UTC" firstStartedPulling="2025-10-01 16:14:46.683946436 +0000 UTC m=+939.585499013" lastFinishedPulling="2025-10-01 16:15:18.983525052 +0000 UTC m=+971.885077629" observedRunningTime="2025-10-01 16:15:23.029657244 +0000 UTC m=+975.931209821" watchObservedRunningTime="2025-10-01 16:15:23.032469836 +0000 UTC m=+975.934022413" Oct 01 16:15:23 crc kubenswrapper[4726]: I1001 16:15:23.413447 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:15:23 crc kubenswrapper[4726]: I1001 16:15:23.413759 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:15:23 crc kubenswrapper[4726]: I1001 16:15:23.413803 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 16:15:23 crc kubenswrapper[4726]: I1001 16:15:23.414474 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3fcc119df2ac4146cc2bd5a3dfb304e41077427fd468db2dfae53b91e3c841a2"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:15:23 crc kubenswrapper[4726]: I1001 16:15:23.414533 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://3fcc119df2ac4146cc2bd5a3dfb304e41077427fd468db2dfae53b91e3c841a2" gracePeriod=600 Oct 01 16:15:23 crc kubenswrapper[4726]: I1001 16:15:23.916867 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75dbebbf-c4f3-4508-901d-05148fd48f74","Type":"ContainerStarted","Data":"cb8f6b26ed1180f41f0273e2461ce4276f12488072d6c566ce51a960f48e6ec2"} Oct 01 16:15:23 crc kubenswrapper[4726]: I1001 16:15:23.918866 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" event={"ID":"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3","Type":"ContainerStarted","Data":"d4f83b310175d8a9f1e888079fe8cd4c37ea7f6bf516e002be0c8502f604e531"} Oct 01 16:15:23 crc kubenswrapper[4726]: I1001 16:15:23.920912 4726 generic.go:334] "Generic (PLEG): container finished" podID="28a68ce6-bbf6-4e46-8414-05ff3ed1a23f" containerID="cf92f6831f92d4e18a0b204c24859bc9026341ad868303d25abe34fb87869188" exitCode=0 Oct 01 16:15:23 crc kubenswrapper[4726]: I1001 16:15:23.921015 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" event={"ID":"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f","Type":"ContainerDied","Data":"cf92f6831f92d4e18a0b204c24859bc9026341ad868303d25abe34fb87869188"} Oct 01 16:15:23 crc kubenswrapper[4726]: I1001 16:15:23.948430 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.270713651 podStartE2EDuration="1m0.948413675s" podCreationTimestamp="2025-10-01 16:14:23 +0000 UTC" firstStartedPulling="2025-10-01 16:14:25.326591355 +0000 UTC m=+918.228143932" lastFinishedPulling="2025-10-01 16:14:47.004291379 +0000 UTC m=+939.905843956" observedRunningTime="2025-10-01 16:15:23.944623625 +0000 UTC m=+976.846176212" watchObservedRunningTime="2025-10-01 16:15:23.948413675 +0000 UTC m=+976.849966252" Oct 01 16:15:24 crc kubenswrapper[4726]: I1001 16:15:24.490783 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:15:24 crc kubenswrapper[4726]: I1001 16:15:24.491149 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:15:24 crc kubenswrapper[4726]: I1001 16:15:24.564178 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:15:24 crc kubenswrapper[4726]: I1001 16:15:24.789431 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 01 16:15:24 crc kubenswrapper[4726]: I1001 16:15:24.934652 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="3fcc119df2ac4146cc2bd5a3dfb304e41077427fd468db2dfae53b91e3c841a2" exitCode=0 Oct 01 16:15:24 crc kubenswrapper[4726]: I1001 16:15:24.934739 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"3fcc119df2ac4146cc2bd5a3dfb304e41077427fd468db2dfae53b91e3c841a2"} Oct 01 16:15:24 crc kubenswrapper[4726]: I1001 16:15:24.936062 4726 scope.go:117] "RemoveContainer" containerID="843d4fde9abf12c4ceb01029d43ed7c5445854afa4d6b76b3edc8ef114cd52e4" Oct 01 16:15:24 crc kubenswrapper[4726]: I1001 16:15:24.936186 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:15:24 crc kubenswrapper[4726]: I1001 16:15:24.936997 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:24 crc kubenswrapper[4726]: I1001 16:15:24.964563 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371974.890236 podStartE2EDuration="1m1.964539414s" podCreationTimestamp="2025-10-01 16:14:23 +0000 UTC" firstStartedPulling="2025-10-01 16:14:25.421621275 +0000 UTC m=+918.323173862" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:15:24.960345912 +0000 UTC m=+977.861898569" watchObservedRunningTime="2025-10-01 16:15:24.964539414 +0000 UTC m=+977.866092011" Oct 01 16:15:24 crc kubenswrapper[4726]: I1001 16:15:24.991536 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" podStartSLOduration=5.991518117 podStartE2EDuration="5.991518117s" podCreationTimestamp="2025-10-01 16:15:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:15:24.988746947 +0000 UTC m=+977.890299524" watchObservedRunningTime="2025-10-01 16:15:24.991518117 +0000 UTC m=+977.893070684" Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.047568 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" podUID="28a68ce6-bbf6-4e46-8414-05ff3ed1a23f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: connect: connection refused" Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.507678 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.576165 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-dns-svc\") pod \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\" (UID: \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\") " Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.576248 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qnsc\" (UniqueName: \"kubernetes.io/projected/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-kube-api-access-4qnsc\") pod \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\" (UID: \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\") " Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.576286 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-config\") pod \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\" (UID: \"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f\") " Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.583329 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-kube-api-access-4qnsc" (OuterVolumeSpecName: "kube-api-access-4qnsc") pod "28a68ce6-bbf6-4e46-8414-05ff3ed1a23f" (UID: "28a68ce6-bbf6-4e46-8414-05ff3ed1a23f"). InnerVolumeSpecName "kube-api-access-4qnsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.622044 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-config" (OuterVolumeSpecName: "config") pod "28a68ce6-bbf6-4e46-8414-05ff3ed1a23f" (UID: "28a68ce6-bbf6-4e46-8414-05ff3ed1a23f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.622083 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "28a68ce6-bbf6-4e46-8414-05ff3ed1a23f" (UID: "28a68ce6-bbf6-4e46-8414-05ff3ed1a23f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.678958 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.679246 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qnsc\" (UniqueName: \"kubernetes.io/projected/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-kube-api-access-4qnsc\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.679345 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.947979 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" event={"ID":"28a68ce6-bbf6-4e46-8414-05ff3ed1a23f","Type":"ContainerDied","Data":"bea3a868171b6db28c225360e932a99e38725d6859f8814b072b20b18b4ae5a6"} Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.948333 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-g4lff" Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.978671 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-g4lff"] Oct 01 16:15:25 crc kubenswrapper[4726]: I1001 16:15:25.991027 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-g4lff"] Oct 01 16:15:26 crc kubenswrapper[4726]: I1001 16:15:26.064249 4726 scope.go:117] "RemoveContainer" containerID="cf92f6831f92d4e18a0b204c24859bc9026341ad868303d25abe34fb87869188" Oct 01 16:15:26 crc kubenswrapper[4726]: I1001 16:15:26.093781 4726 scope.go:117] "RemoveContainer" containerID="e7c8114eb0341cbd293f2d63b089410bdc6432b67ad17ced326d0337f00ad3c1" Oct 01 16:15:27 crc kubenswrapper[4726]: I1001 16:15:27.819579 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28a68ce6-bbf6-4e46-8414-05ff3ed1a23f" path="/var/lib/kubelet/pods/28a68ce6-bbf6-4e46-8414-05ff3ed1a23f/volumes" Oct 01 16:15:27 crc kubenswrapper[4726]: I1001 16:15:27.986773 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"db368996f7ef78489a378108debc56732f3e0a06eb79040d4b7e667cfd8503a8"} Oct 01 16:15:28 crc kubenswrapper[4726]: I1001 16:15:28.539668 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 01 16:15:28 crc kubenswrapper[4726]: I1001 16:15:28.612837 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 01 16:15:29 crc kubenswrapper[4726]: I1001 16:15:29.385135 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 01 16:15:29 crc kubenswrapper[4726]: I1001 16:15:29.652040 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:15:29 crc kubenswrapper[4726]: I1001 16:15:29.730424 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qsbql"] Oct 01 16:15:29 crc kubenswrapper[4726]: I1001 16:15:29.733510 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-qsbql" podUID="47538946-adc6-42ea-a62a-3cf2fb6a3206" containerName="dnsmasq-dns" containerID="cri-o://39fcd9cffb0b4be1dabd72b8d8c128df909bc9caeff126760a573dda836b0bfc" gracePeriod=10 Oct 01 16:15:30 crc kubenswrapper[4726]: I1001 16:15:30.004133 4726 generic.go:334] "Generic (PLEG): container finished" podID="47538946-adc6-42ea-a62a-3cf2fb6a3206" containerID="39fcd9cffb0b4be1dabd72b8d8c128df909bc9caeff126760a573dda836b0bfc" exitCode=0 Oct 01 16:15:30 crc kubenswrapper[4726]: I1001 16:15:30.004183 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qsbql" event={"ID":"47538946-adc6-42ea-a62a-3cf2fb6a3206","Type":"ContainerDied","Data":"39fcd9cffb0b4be1dabd72b8d8c128df909bc9caeff126760a573dda836b0bfc"} Oct 01 16:15:30 crc kubenswrapper[4726]: I1001 16:15:30.527326 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 01 16:15:30 crc kubenswrapper[4726]: I1001 16:15:30.587476 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="d743678c-e92d-4515-88d0-27e4a5086aac" containerName="galera" probeResult="failure" output=< Oct 01 16:15:30 crc kubenswrapper[4726]: wsrep_local_state_comment (Joined) differs from Synced Oct 01 16:15:30 crc kubenswrapper[4726]: > Oct 01 16:15:32 crc kubenswrapper[4726]: I1001 16:15:32.502973 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:15:32 crc kubenswrapper[4726]: E1001 16:15:32.503293 4726 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 16:15:32 crc kubenswrapper[4726]: E1001 16:15:32.503728 4726 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 16:15:32 crc kubenswrapper[4726]: E1001 16:15:32.503819 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift podName:6d192df1-7269-4493-b17b-4037addc226e nodeName:}" failed. No retries permitted until 2025-10-01 16:16:04.503788899 +0000 UTC m=+1017.405341506 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift") pod "swift-storage-0" (UID: "6d192df1-7269-4493-b17b-4037addc226e") : configmap "swift-ring-files" not found Oct 01 16:15:32 crc kubenswrapper[4726]: I1001 16:15:32.637084 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:15:32 crc kubenswrapper[4726]: I1001 16:15:32.706607 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5spq9\" (UniqueName: \"kubernetes.io/projected/47538946-adc6-42ea-a62a-3cf2fb6a3206-kube-api-access-5spq9\") pod \"47538946-adc6-42ea-a62a-3cf2fb6a3206\" (UID: \"47538946-adc6-42ea-a62a-3cf2fb6a3206\") " Oct 01 16:15:32 crc kubenswrapper[4726]: I1001 16:15:32.706682 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47538946-adc6-42ea-a62a-3cf2fb6a3206-dns-svc\") pod \"47538946-adc6-42ea-a62a-3cf2fb6a3206\" (UID: \"47538946-adc6-42ea-a62a-3cf2fb6a3206\") " Oct 01 16:15:32 crc kubenswrapper[4726]: I1001 16:15:32.706730 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47538946-adc6-42ea-a62a-3cf2fb6a3206-config\") pod \"47538946-adc6-42ea-a62a-3cf2fb6a3206\" (UID: \"47538946-adc6-42ea-a62a-3cf2fb6a3206\") " Oct 01 16:15:32 crc kubenswrapper[4726]: I1001 16:15:32.729385 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47538946-adc6-42ea-a62a-3cf2fb6a3206-kube-api-access-5spq9" (OuterVolumeSpecName: "kube-api-access-5spq9") pod "47538946-adc6-42ea-a62a-3cf2fb6a3206" (UID: "47538946-adc6-42ea-a62a-3cf2fb6a3206"). InnerVolumeSpecName "kube-api-access-5spq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:32 crc kubenswrapper[4726]: I1001 16:15:32.772511 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47538946-adc6-42ea-a62a-3cf2fb6a3206-config" (OuterVolumeSpecName: "config") pod "47538946-adc6-42ea-a62a-3cf2fb6a3206" (UID: "47538946-adc6-42ea-a62a-3cf2fb6a3206"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:32 crc kubenswrapper[4726]: I1001 16:15:32.776256 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47538946-adc6-42ea-a62a-3cf2fb6a3206-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "47538946-adc6-42ea-a62a-3cf2fb6a3206" (UID: "47538946-adc6-42ea-a62a-3cf2fb6a3206"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:32 crc kubenswrapper[4726]: I1001 16:15:32.809301 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5spq9\" (UniqueName: \"kubernetes.io/projected/47538946-adc6-42ea-a62a-3cf2fb6a3206-kube-api-access-5spq9\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:32 crc kubenswrapper[4726]: I1001 16:15:32.809329 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47538946-adc6-42ea-a62a-3cf2fb6a3206-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:32 crc kubenswrapper[4726]: I1001 16:15:32.809338 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47538946-adc6-42ea-a62a-3cf2fb6a3206-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:32 crc kubenswrapper[4726]: I1001 16:15:32.869357 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-z4lk4" podUID="f67ec6c1-adf7-47f9-855f-fde38e0430c7" containerName="ovn-controller" probeResult="failure" output=< Oct 01 16:15:32 crc kubenswrapper[4726]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 01 16:15:32 crc kubenswrapper[4726]: > Oct 01 16:15:33 crc kubenswrapper[4726]: I1001 16:15:33.026691 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qsbql" event={"ID":"47538946-adc6-42ea-a62a-3cf2fb6a3206","Type":"ContainerDied","Data":"890ca581c9fae19ca9932f4aa3f240b86bf7352ad6a3d426290fb4be7540e273"} Oct 01 16:15:33 crc kubenswrapper[4726]: I1001 16:15:33.026750 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qsbql" Oct 01 16:15:33 crc kubenswrapper[4726]: I1001 16:15:33.026774 4726 scope.go:117] "RemoveContainer" containerID="39fcd9cffb0b4be1dabd72b8d8c128df909bc9caeff126760a573dda836b0bfc" Oct 01 16:15:33 crc kubenswrapper[4726]: I1001 16:15:33.088594 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qsbql"] Oct 01 16:15:33 crc kubenswrapper[4726]: I1001 16:15:33.092803 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qsbql"] Oct 01 16:15:33 crc kubenswrapper[4726]: I1001 16:15:33.827507 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47538946-adc6-42ea-a62a-3cf2fb6a3206" path="/var/lib/kubelet/pods/47538946-adc6-42ea-a62a-3cf2fb6a3206/volumes" Oct 01 16:15:34 crc kubenswrapper[4726]: I1001 16:15:34.205593 4726 scope.go:117] "RemoveContainer" containerID="4c701161a5ed766c6ce3aff83c08c52cad5ac6acf4bbf04ff0440cb72f53a473" Oct 01 16:15:34 crc kubenswrapper[4726]: I1001 16:15:34.553542 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:15:34 crc kubenswrapper[4726]: I1001 16:15:34.596608 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jnxwz"] Oct 01 16:15:34 crc kubenswrapper[4726]: I1001 16:15:34.790557 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="57593043-8bb5-4967-8a6c-1a55a3e8894c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Oct 01 16:15:34 crc kubenswrapper[4726]: I1001 16:15:34.837460 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="75dbebbf-c4f3-4508-901d-05148fd48f74" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.044116 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"23600bcf-22d0-4de1-be32-5adf413748b4","Type":"ContainerStarted","Data":"4800ee4f9fb6ba5489c6cd4e59480ad1c79effbb3b208ac3cb8c735e82d528d3"} Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.049752 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t4hqx" event={"ID":"4ab0d371-af59-407e-b596-3c533ad031b8","Type":"ContainerStarted","Data":"3443129e3e0ade132d2f07767b542038259f11fa4d00f1cd2a206c9190d85c67"} Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.052979 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab","Type":"ContainerStarted","Data":"aa5bfa28f55bb141861796d0ac7a0adbce9ac01f964d19e10c64f68fc25c490a"} Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.056213 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-h598x" event={"ID":"72fa255e-484d-43ab-a837-d31c20e8e4d4","Type":"ContainerStarted","Data":"43e26afe4924cd8e515121930c4a984e97401c91a0739171cdf0e8428ee329a6"} Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.056253 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jnxwz" podUID="12562596-fc27-4e0f-84b1-f06ee267cbe5" containerName="registry-server" containerID="cri-o://7354a0ba99e8490acd9a196a901a72214e653dc109e54f6e71266f2b2cda33f6" gracePeriod=2 Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.066663 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=13.323124868 podStartE2EDuration="1m0.066641095s" podCreationTimestamp="2025-10-01 16:14:35 +0000 UTC" firstStartedPulling="2025-10-01 16:14:47.462832084 +0000 UTC m=+940.364384661" lastFinishedPulling="2025-10-01 16:15:34.206348261 +0000 UTC m=+987.107900888" observedRunningTime="2025-10-01 16:15:35.064632077 +0000 UTC m=+987.966184674" watchObservedRunningTime="2025-10-01 16:15:35.066641095 +0000 UTC m=+987.968193672" Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.103027 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-t4hqx" podStartSLOduration=11.270711052 podStartE2EDuration="31.103005871s" podCreationTimestamp="2025-10-01 16:15:04 +0000 UTC" firstStartedPulling="2025-10-01 16:15:14.387175533 +0000 UTC m=+967.288728100" lastFinishedPulling="2025-10-01 16:15:34.219470332 +0000 UTC m=+987.121022919" observedRunningTime="2025-10-01 16:15:35.091802086 +0000 UTC m=+987.993354683" watchObservedRunningTime="2025-10-01 16:15:35.103005871 +0000 UTC m=+988.004558458" Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.117308 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=16.504373381 podStartE2EDuration="1m3.117286516s" podCreationTimestamp="2025-10-01 16:14:32 +0000 UTC" firstStartedPulling="2025-10-01 16:14:47.592891401 +0000 UTC m=+940.494443978" lastFinishedPulling="2025-10-01 16:15:34.205804496 +0000 UTC m=+987.107357113" observedRunningTime="2025-10-01 16:15:35.115647268 +0000 UTC m=+988.017199855" watchObservedRunningTime="2025-10-01 16:15:35.117286516 +0000 UTC m=+988.018839093" Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.140597 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-h598x" podStartSLOduration=2.712559664 podStartE2EDuration="17.140572072s" podCreationTimestamp="2025-10-01 16:15:18 +0000 UTC" firstStartedPulling="2025-10-01 16:15:19.777365195 +0000 UTC m=+972.678917782" lastFinishedPulling="2025-10-01 16:15:34.205377603 +0000 UTC m=+987.106930190" observedRunningTime="2025-10-01 16:15:35.13464805 +0000 UTC m=+988.036200697" watchObservedRunningTime="2025-10-01 16:15:35.140572072 +0000 UTC m=+988.042124649" Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.503356 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.556501 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12562596-fc27-4e0f-84b1-f06ee267cbe5-utilities\") pod \"12562596-fc27-4e0f-84b1-f06ee267cbe5\" (UID: \"12562596-fc27-4e0f-84b1-f06ee267cbe5\") " Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.556636 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12562596-fc27-4e0f-84b1-f06ee267cbe5-catalog-content\") pod \"12562596-fc27-4e0f-84b1-f06ee267cbe5\" (UID: \"12562596-fc27-4e0f-84b1-f06ee267cbe5\") " Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.556676 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hlgm\" (UniqueName: \"kubernetes.io/projected/12562596-fc27-4e0f-84b1-f06ee267cbe5-kube-api-access-4hlgm\") pod \"12562596-fc27-4e0f-84b1-f06ee267cbe5\" (UID: \"12562596-fc27-4e0f-84b1-f06ee267cbe5\") " Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.557395 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12562596-fc27-4e0f-84b1-f06ee267cbe5-utilities" (OuterVolumeSpecName: "utilities") pod "12562596-fc27-4e0f-84b1-f06ee267cbe5" (UID: "12562596-fc27-4e0f-84b1-f06ee267cbe5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.564414 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12562596-fc27-4e0f-84b1-f06ee267cbe5-kube-api-access-4hlgm" (OuterVolumeSpecName: "kube-api-access-4hlgm") pod "12562596-fc27-4e0f-84b1-f06ee267cbe5" (UID: "12562596-fc27-4e0f-84b1-f06ee267cbe5"). InnerVolumeSpecName "kube-api-access-4hlgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.606912 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12562596-fc27-4e0f-84b1-f06ee267cbe5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "12562596-fc27-4e0f-84b1-f06ee267cbe5" (UID: "12562596-fc27-4e0f-84b1-f06ee267cbe5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.659389 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12562596-fc27-4e0f-84b1-f06ee267cbe5-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.659454 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12562596-fc27-4e0f-84b1-f06ee267cbe5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:35 crc kubenswrapper[4726]: I1001 16:15:35.659475 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hlgm\" (UniqueName: \"kubernetes.io/projected/12562596-fc27-4e0f-84b1-f06ee267cbe5-kube-api-access-4hlgm\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.065570 4726 generic.go:334] "Generic (PLEG): container finished" podID="12562596-fc27-4e0f-84b1-f06ee267cbe5" containerID="7354a0ba99e8490acd9a196a901a72214e653dc109e54f6e71266f2b2cda33f6" exitCode=0 Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.065691 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnxwz" event={"ID":"12562596-fc27-4e0f-84b1-f06ee267cbe5","Type":"ContainerDied","Data":"7354a0ba99e8490acd9a196a901a72214e653dc109e54f6e71266f2b2cda33f6"} Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.065737 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnxwz" Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.065765 4726 scope.go:117] "RemoveContainer" containerID="7354a0ba99e8490acd9a196a901a72214e653dc109e54f6e71266f2b2cda33f6" Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.065747 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnxwz" event={"ID":"12562596-fc27-4e0f-84b1-f06ee267cbe5","Type":"ContainerDied","Data":"23f9158931427d4720b9f6e2a175452294ecf324f57b98858a48051a2bde3f73"} Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.106422 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jnxwz"] Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.123443 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jnxwz"] Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.164141 4726 scope.go:117] "RemoveContainer" containerID="db4e3de37114da59675196f0388b9b45602f37d30d92e8683923aed099b5e05f" Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.187514 4726 scope.go:117] "RemoveContainer" containerID="0c19bbd94879716e829aa57d6a696869d19a0232c791904e60388fcd0ef129cc" Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.231823 4726 scope.go:117] "RemoveContainer" containerID="7354a0ba99e8490acd9a196a901a72214e653dc109e54f6e71266f2b2cda33f6" Oct 01 16:15:36 crc kubenswrapper[4726]: E1001 16:15:36.232299 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7354a0ba99e8490acd9a196a901a72214e653dc109e54f6e71266f2b2cda33f6\": container with ID starting with 7354a0ba99e8490acd9a196a901a72214e653dc109e54f6e71266f2b2cda33f6 not found: ID does not exist" containerID="7354a0ba99e8490acd9a196a901a72214e653dc109e54f6e71266f2b2cda33f6" Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.232329 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7354a0ba99e8490acd9a196a901a72214e653dc109e54f6e71266f2b2cda33f6"} err="failed to get container status \"7354a0ba99e8490acd9a196a901a72214e653dc109e54f6e71266f2b2cda33f6\": rpc error: code = NotFound desc = could not find container \"7354a0ba99e8490acd9a196a901a72214e653dc109e54f6e71266f2b2cda33f6\": container with ID starting with 7354a0ba99e8490acd9a196a901a72214e653dc109e54f6e71266f2b2cda33f6 not found: ID does not exist" Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.232352 4726 scope.go:117] "RemoveContainer" containerID="db4e3de37114da59675196f0388b9b45602f37d30d92e8683923aed099b5e05f" Oct 01 16:15:36 crc kubenswrapper[4726]: E1001 16:15:36.232771 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db4e3de37114da59675196f0388b9b45602f37d30d92e8683923aed099b5e05f\": container with ID starting with db4e3de37114da59675196f0388b9b45602f37d30d92e8683923aed099b5e05f not found: ID does not exist" containerID="db4e3de37114da59675196f0388b9b45602f37d30d92e8683923aed099b5e05f" Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.232835 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4e3de37114da59675196f0388b9b45602f37d30d92e8683923aed099b5e05f"} err="failed to get container status \"db4e3de37114da59675196f0388b9b45602f37d30d92e8683923aed099b5e05f\": rpc error: code = NotFound desc = could not find container \"db4e3de37114da59675196f0388b9b45602f37d30d92e8683923aed099b5e05f\": container with ID starting with db4e3de37114da59675196f0388b9b45602f37d30d92e8683923aed099b5e05f not found: ID does not exist" Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.232871 4726 scope.go:117] "RemoveContainer" containerID="0c19bbd94879716e829aa57d6a696869d19a0232c791904e60388fcd0ef129cc" Oct 01 16:15:36 crc kubenswrapper[4726]: E1001 16:15:36.233376 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c19bbd94879716e829aa57d6a696869d19a0232c791904e60388fcd0ef129cc\": container with ID starting with 0c19bbd94879716e829aa57d6a696869d19a0232c791904e60388fcd0ef129cc not found: ID does not exist" containerID="0c19bbd94879716e829aa57d6a696869d19a0232c791904e60388fcd0ef129cc" Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.233425 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c19bbd94879716e829aa57d6a696869d19a0232c791904e60388fcd0ef129cc"} err="failed to get container status \"0c19bbd94879716e829aa57d6a696869d19a0232c791904e60388fcd0ef129cc\": rpc error: code = NotFound desc = could not find container \"0c19bbd94879716e829aa57d6a696869d19a0232c791904e60388fcd0ef129cc\": container with ID starting with 0c19bbd94879716e829aa57d6a696869d19a0232c791904e60388fcd0ef129cc not found: ID does not exist" Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.379250 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.440418 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 01 16:15:36 crc kubenswrapper[4726]: I1001 16:15:36.505555 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.075993 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.135620 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.251680 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-b62tk"] Oct 01 16:15:37 crc kubenswrapper[4726]: E1001 16:15:37.252077 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47538946-adc6-42ea-a62a-3cf2fb6a3206" containerName="init" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.252093 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="47538946-adc6-42ea-a62a-3cf2fb6a3206" containerName="init" Oct 01 16:15:37 crc kubenswrapper[4726]: E1001 16:15:37.252105 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12562596-fc27-4e0f-84b1-f06ee267cbe5" containerName="extract-content" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.252112 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="12562596-fc27-4e0f-84b1-f06ee267cbe5" containerName="extract-content" Oct 01 16:15:37 crc kubenswrapper[4726]: E1001 16:15:37.252124 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47538946-adc6-42ea-a62a-3cf2fb6a3206" containerName="dnsmasq-dns" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.252130 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="47538946-adc6-42ea-a62a-3cf2fb6a3206" containerName="dnsmasq-dns" Oct 01 16:15:37 crc kubenswrapper[4726]: E1001 16:15:37.252149 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a68ce6-bbf6-4e46-8414-05ff3ed1a23f" containerName="dnsmasq-dns" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.252155 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a68ce6-bbf6-4e46-8414-05ff3ed1a23f" containerName="dnsmasq-dns" Oct 01 16:15:37 crc kubenswrapper[4726]: E1001 16:15:37.252166 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a68ce6-bbf6-4e46-8414-05ff3ed1a23f" containerName="init" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.252171 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a68ce6-bbf6-4e46-8414-05ff3ed1a23f" containerName="init" Oct 01 16:15:37 crc kubenswrapper[4726]: E1001 16:15:37.252186 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12562596-fc27-4e0f-84b1-f06ee267cbe5" containerName="registry-server" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.252192 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="12562596-fc27-4e0f-84b1-f06ee267cbe5" containerName="registry-server" Oct 01 16:15:37 crc kubenswrapper[4726]: E1001 16:15:37.252201 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12562596-fc27-4e0f-84b1-f06ee267cbe5" containerName="extract-utilities" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.252206 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="12562596-fc27-4e0f-84b1-f06ee267cbe5" containerName="extract-utilities" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.252351 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="28a68ce6-bbf6-4e46-8414-05ff3ed1a23f" containerName="dnsmasq-dns" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.252363 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="12562596-fc27-4e0f-84b1-f06ee267cbe5" containerName="registry-server" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.252373 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="47538946-adc6-42ea-a62a-3cf2fb6a3206" containerName="dnsmasq-dns" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.252898 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-b62tk" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.258706 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-b62tk"] Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.393677 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6f2m\" (UniqueName: \"kubernetes.io/projected/15a72138-dafd-48f5-95cb-420e35e00609-kube-api-access-x6f2m\") pod \"keystone-db-create-b62tk\" (UID: \"15a72138-dafd-48f5-95cb-420e35e00609\") " pod="openstack/keystone-db-create-b62tk" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.454636 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-gcts2"] Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.455981 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gcts2" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.468425 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-gcts2"] Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.494878 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6f2m\" (UniqueName: \"kubernetes.io/projected/15a72138-dafd-48f5-95cb-420e35e00609-kube-api-access-x6f2m\") pod \"keystone-db-create-b62tk\" (UID: \"15a72138-dafd-48f5-95cb-420e35e00609\") " pod="openstack/keystone-db-create-b62tk" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.521261 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6f2m\" (UniqueName: \"kubernetes.io/projected/15a72138-dafd-48f5-95cb-420e35e00609-kube-api-access-x6f2m\") pod \"keystone-db-create-b62tk\" (UID: \"15a72138-dafd-48f5-95cb-420e35e00609\") " pod="openstack/keystone-db-create-b62tk" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.596771 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f67l\" (UniqueName: \"kubernetes.io/projected/3e34fc66-c06d-4ed5-96f1-f54dc1fe4765-kube-api-access-5f67l\") pod \"placement-db-create-gcts2\" (UID: \"3e34fc66-c06d-4ed5-96f1-f54dc1fe4765\") " pod="openstack/placement-db-create-gcts2" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.604119 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-b62tk" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.700266 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f67l\" (UniqueName: \"kubernetes.io/projected/3e34fc66-c06d-4ed5-96f1-f54dc1fe4765-kube-api-access-5f67l\") pod \"placement-db-create-gcts2\" (UID: \"3e34fc66-c06d-4ed5-96f1-f54dc1fe4765\") " pod="openstack/placement-db-create-gcts2" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.727858 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f67l\" (UniqueName: \"kubernetes.io/projected/3e34fc66-c06d-4ed5-96f1-f54dc1fe4765-kube-api-access-5f67l\") pod \"placement-db-create-gcts2\" (UID: \"3e34fc66-c06d-4ed5-96f1-f54dc1fe4765\") " pod="openstack/placement-db-create-gcts2" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.775537 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gcts2" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.819603 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12562596-fc27-4e0f-84b1-f06ee267cbe5" path="/var/lib/kubelet/pods/12562596-fc27-4e0f-84b1-f06ee267cbe5/volumes" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.821020 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.821194 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.858160 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.892263 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-z4lk4" podUID="f67ec6c1-adf7-47f9-855f-fde38e0430c7" containerName="ovn-controller" probeResult="failure" output=< Oct 01 16:15:37 crc kubenswrapper[4726]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 01 16:15:37 crc kubenswrapper[4726]: > Oct 01 16:15:37 crc kubenswrapper[4726]: I1001 16:15:37.932257 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:15:38 crc kubenswrapper[4726]: W1001 16:15:38.057955 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15a72138_dafd_48f5_95cb_420e35e00609.slice/crio-76edfc443b52627280a320f87effd8cd839bc23196abdb16dea2e9f469c49189 WatchSource:0}: Error finding container 76edfc443b52627280a320f87effd8cd839bc23196abdb16dea2e9f469c49189: Status 404 returned error can't find the container with id 76edfc443b52627280a320f87effd8cd839bc23196abdb16dea2e9f469c49189 Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.061457 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-b62tk"] Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.082827 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-b62tk" event={"ID":"15a72138-dafd-48f5-95cb-420e35e00609","Type":"ContainerStarted","Data":"76edfc443b52627280a320f87effd8cd839bc23196abdb16dea2e9f469c49189"} Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.144082 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.214860 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-gcts2"] Oct 01 16:15:38 crc kubenswrapper[4726]: W1001 16:15:38.220012 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e34fc66_c06d_4ed5_96f1_f54dc1fe4765.slice/crio-77794423a5f354847194cc727503a0c6dbd58a3a72af962a52a093730b2ae2fa WatchSource:0}: Error finding container 77794423a5f354847194cc727503a0c6dbd58a3a72af962a52a093730b2ae2fa: Status 404 returned error can't find the container with id 77794423a5f354847194cc727503a0c6dbd58a3a72af962a52a093730b2ae2fa Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.301177 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.304793 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.309390 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.309583 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.309745 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-2zcxt" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.315151 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.331621 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.422025 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb9896de-b781-451d-8207-093218e40684-config\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.422137 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9896de-b781-451d-8207-093218e40684-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.422227 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79kkt\" (UniqueName: \"kubernetes.io/projected/cb9896de-b781-451d-8207-093218e40684-kube-api-access-79kkt\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.422305 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9896de-b781-451d-8207-093218e40684-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.422369 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb9896de-b781-451d-8207-093218e40684-scripts\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.422447 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9896de-b781-451d-8207-093218e40684-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.422526 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cb9896de-b781-451d-8207-093218e40684-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.524435 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb9896de-b781-451d-8207-093218e40684-config\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.524786 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9896de-b781-451d-8207-093218e40684-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.524832 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79kkt\" (UniqueName: \"kubernetes.io/projected/cb9896de-b781-451d-8207-093218e40684-kube-api-access-79kkt\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.524887 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9896de-b781-451d-8207-093218e40684-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.524913 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb9896de-b781-451d-8207-093218e40684-scripts\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.524935 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9896de-b781-451d-8207-093218e40684-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.524973 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cb9896de-b781-451d-8207-093218e40684-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.525626 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cb9896de-b781-451d-8207-093218e40684-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.527625 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb9896de-b781-451d-8207-093218e40684-config\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.529562 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb9896de-b781-451d-8207-093218e40684-scripts\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.531695 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9896de-b781-451d-8207-093218e40684-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.532012 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9896de-b781-451d-8207-093218e40684-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.534181 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9896de-b781-451d-8207-093218e40684-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.541364 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79kkt\" (UniqueName: \"kubernetes.io/projected/cb9896de-b781-451d-8207-093218e40684-kube-api-access-79kkt\") pod \"ovn-northd-0\" (UID: \"cb9896de-b781-451d-8207-093218e40684\") " pod="openstack/ovn-northd-0" Oct 01 16:15:38 crc kubenswrapper[4726]: I1001 16:15:38.687211 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 01 16:15:39 crc kubenswrapper[4726]: I1001 16:15:39.091455 4726 generic.go:334] "Generic (PLEG): container finished" podID="15a72138-dafd-48f5-95cb-420e35e00609" containerID="c87bf65c6246937eb2398b3c487d03304986c354f5e8f24957cc8aa27aa5b7d8" exitCode=0 Oct 01 16:15:39 crc kubenswrapper[4726]: I1001 16:15:39.091831 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-b62tk" event={"ID":"15a72138-dafd-48f5-95cb-420e35e00609","Type":"ContainerDied","Data":"c87bf65c6246937eb2398b3c487d03304986c354f5e8f24957cc8aa27aa5b7d8"} Oct 01 16:15:39 crc kubenswrapper[4726]: I1001 16:15:39.093662 4726 generic.go:334] "Generic (PLEG): container finished" podID="3e34fc66-c06d-4ed5-96f1-f54dc1fe4765" containerID="a9a17ae1b6d831f93318caa0fc9fd8249b771d2b7792ffa6ff4d089140f40ee8" exitCode=0 Oct 01 16:15:39 crc kubenswrapper[4726]: I1001 16:15:39.093693 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-gcts2" event={"ID":"3e34fc66-c06d-4ed5-96f1-f54dc1fe4765","Type":"ContainerDied","Data":"a9a17ae1b6d831f93318caa0fc9fd8249b771d2b7792ffa6ff4d089140f40ee8"} Oct 01 16:15:39 crc kubenswrapper[4726]: I1001 16:15:39.093729 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-gcts2" event={"ID":"3e34fc66-c06d-4ed5-96f1-f54dc1fe4765","Type":"ContainerStarted","Data":"77794423a5f354847194cc727503a0c6dbd58a3a72af962a52a093730b2ae2fa"} Oct 01 16:15:39 crc kubenswrapper[4726]: I1001 16:15:39.126623 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 01 16:15:39 crc kubenswrapper[4726]: W1001 16:15:39.135381 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb9896de_b781_451d_8207_093218e40684.slice/crio-83542ac066e20c21227de9e7bf22c34b5df7f5241eeace88e217cae2385b886a WatchSource:0}: Error finding container 83542ac066e20c21227de9e7bf22c34b5df7f5241eeace88e217cae2385b886a: Status 404 returned error can't find the container with id 83542ac066e20c21227de9e7bf22c34b5df7f5241eeace88e217cae2385b886a Oct 01 16:15:40 crc kubenswrapper[4726]: I1001 16:15:40.105583 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"cb9896de-b781-451d-8207-093218e40684","Type":"ContainerStarted","Data":"83542ac066e20c21227de9e7bf22c34b5df7f5241eeace88e217cae2385b886a"} Oct 01 16:15:40 crc kubenswrapper[4726]: I1001 16:15:40.511348 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gcts2" Oct 01 16:15:40 crc kubenswrapper[4726]: I1001 16:15:40.519040 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-b62tk" Oct 01 16:15:40 crc kubenswrapper[4726]: I1001 16:15:40.658874 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5f67l\" (UniqueName: \"kubernetes.io/projected/3e34fc66-c06d-4ed5-96f1-f54dc1fe4765-kube-api-access-5f67l\") pod \"3e34fc66-c06d-4ed5-96f1-f54dc1fe4765\" (UID: \"3e34fc66-c06d-4ed5-96f1-f54dc1fe4765\") " Oct 01 16:15:40 crc kubenswrapper[4726]: I1001 16:15:40.658967 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6f2m\" (UniqueName: \"kubernetes.io/projected/15a72138-dafd-48f5-95cb-420e35e00609-kube-api-access-x6f2m\") pod \"15a72138-dafd-48f5-95cb-420e35e00609\" (UID: \"15a72138-dafd-48f5-95cb-420e35e00609\") " Oct 01 16:15:40 crc kubenswrapper[4726]: I1001 16:15:40.665442 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15a72138-dafd-48f5-95cb-420e35e00609-kube-api-access-x6f2m" (OuterVolumeSpecName: "kube-api-access-x6f2m") pod "15a72138-dafd-48f5-95cb-420e35e00609" (UID: "15a72138-dafd-48f5-95cb-420e35e00609"). InnerVolumeSpecName "kube-api-access-x6f2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:40 crc kubenswrapper[4726]: I1001 16:15:40.666265 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e34fc66-c06d-4ed5-96f1-f54dc1fe4765-kube-api-access-5f67l" (OuterVolumeSpecName: "kube-api-access-5f67l") pod "3e34fc66-c06d-4ed5-96f1-f54dc1fe4765" (UID: "3e34fc66-c06d-4ed5-96f1-f54dc1fe4765"). InnerVolumeSpecName "kube-api-access-5f67l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:40 crc kubenswrapper[4726]: I1001 16:15:40.761223 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6f2m\" (UniqueName: \"kubernetes.io/projected/15a72138-dafd-48f5-95cb-420e35e00609-kube-api-access-x6f2m\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:40 crc kubenswrapper[4726]: I1001 16:15:40.761261 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5f67l\" (UniqueName: \"kubernetes.io/projected/3e34fc66-c06d-4ed5-96f1-f54dc1fe4765-kube-api-access-5f67l\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:41 crc kubenswrapper[4726]: I1001 16:15:41.113992 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-b62tk" Oct 01 16:15:41 crc kubenswrapper[4726]: I1001 16:15:41.115176 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-b62tk" event={"ID":"15a72138-dafd-48f5-95cb-420e35e00609","Type":"ContainerDied","Data":"76edfc443b52627280a320f87effd8cd839bc23196abdb16dea2e9f469c49189"} Oct 01 16:15:41 crc kubenswrapper[4726]: I1001 16:15:41.115302 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76edfc443b52627280a320f87effd8cd839bc23196abdb16dea2e9f469c49189" Oct 01 16:15:41 crc kubenswrapper[4726]: I1001 16:15:41.116692 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-gcts2" event={"ID":"3e34fc66-c06d-4ed5-96f1-f54dc1fe4765","Type":"ContainerDied","Data":"77794423a5f354847194cc727503a0c6dbd58a3a72af962a52a093730b2ae2fa"} Oct 01 16:15:41 crc kubenswrapper[4726]: I1001 16:15:41.116833 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77794423a5f354847194cc727503a0c6dbd58a3a72af962a52a093730b2ae2fa" Oct 01 16:15:41 crc kubenswrapper[4726]: I1001 16:15:41.117024 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gcts2" Oct 01 16:15:42 crc kubenswrapper[4726]: I1001 16:15:42.642809 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-gj94c"] Oct 01 16:15:42 crc kubenswrapper[4726]: E1001 16:15:42.643843 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e34fc66-c06d-4ed5-96f1-f54dc1fe4765" containerName="mariadb-database-create" Oct 01 16:15:42 crc kubenswrapper[4726]: I1001 16:15:42.643858 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e34fc66-c06d-4ed5-96f1-f54dc1fe4765" containerName="mariadb-database-create" Oct 01 16:15:42 crc kubenswrapper[4726]: E1001 16:15:42.643884 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15a72138-dafd-48f5-95cb-420e35e00609" containerName="mariadb-database-create" Oct 01 16:15:42 crc kubenswrapper[4726]: I1001 16:15:42.643890 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="15a72138-dafd-48f5-95cb-420e35e00609" containerName="mariadb-database-create" Oct 01 16:15:42 crc kubenswrapper[4726]: I1001 16:15:42.644068 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e34fc66-c06d-4ed5-96f1-f54dc1fe4765" containerName="mariadb-database-create" Oct 01 16:15:42 crc kubenswrapper[4726]: I1001 16:15:42.644098 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="15a72138-dafd-48f5-95cb-420e35e00609" containerName="mariadb-database-create" Oct 01 16:15:42 crc kubenswrapper[4726]: I1001 16:15:42.644713 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gj94c" Oct 01 16:15:42 crc kubenswrapper[4726]: I1001 16:15:42.650711 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-gj94c"] Oct 01 16:15:42 crc kubenswrapper[4726]: I1001 16:15:42.797131 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58knx\" (UniqueName: \"kubernetes.io/projected/de31038e-fe8f-4d23-b875-aad3c596fbea-kube-api-access-58knx\") pod \"glance-db-create-gj94c\" (UID: \"de31038e-fe8f-4d23-b875-aad3c596fbea\") " pod="openstack/glance-db-create-gj94c" Oct 01 16:15:42 crc kubenswrapper[4726]: I1001 16:15:42.864461 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-z4lk4" podUID="f67ec6c1-adf7-47f9-855f-fde38e0430c7" containerName="ovn-controller" probeResult="failure" output=< Oct 01 16:15:42 crc kubenswrapper[4726]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 01 16:15:42 crc kubenswrapper[4726]: > Oct 01 16:15:42 crc kubenswrapper[4726]: I1001 16:15:42.898847 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58knx\" (UniqueName: \"kubernetes.io/projected/de31038e-fe8f-4d23-b875-aad3c596fbea-kube-api-access-58knx\") pod \"glance-db-create-gj94c\" (UID: \"de31038e-fe8f-4d23-b875-aad3c596fbea\") " pod="openstack/glance-db-create-gj94c" Oct 01 16:15:42 crc kubenswrapper[4726]: I1001 16:15:42.916962 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58knx\" (UniqueName: \"kubernetes.io/projected/de31038e-fe8f-4d23-b875-aad3c596fbea-kube-api-access-58knx\") pod \"glance-db-create-gj94c\" (UID: \"de31038e-fe8f-4d23-b875-aad3c596fbea\") " pod="openstack/glance-db-create-gj94c" Oct 01 16:15:42 crc kubenswrapper[4726]: I1001 16:15:42.961380 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gj94c" Oct 01 16:15:43 crc kubenswrapper[4726]: I1001 16:15:43.138643 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"cb9896de-b781-451d-8207-093218e40684","Type":"ContainerStarted","Data":"e261a8312228eeb88912e4e8b96c5e94b90dffab220203c3bf0a7ab948a6aebd"} Oct 01 16:15:43 crc kubenswrapper[4726]: I1001 16:15:43.138969 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"cb9896de-b781-451d-8207-093218e40684","Type":"ContainerStarted","Data":"e67bd7360346d3ee7ee12cb73de8de578838fa414e9b5c5c8d3e3d574d094600"} Oct 01 16:15:43 crc kubenswrapper[4726]: I1001 16:15:43.139016 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 01 16:15:43 crc kubenswrapper[4726]: I1001 16:15:43.141317 4726 generic.go:334] "Generic (PLEG): container finished" podID="4ab0d371-af59-407e-b596-3c533ad031b8" containerID="3443129e3e0ade132d2f07767b542038259f11fa4d00f1cd2a206c9190d85c67" exitCode=0 Oct 01 16:15:43 crc kubenswrapper[4726]: I1001 16:15:43.141359 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t4hqx" event={"ID":"4ab0d371-af59-407e-b596-3c533ad031b8","Type":"ContainerDied","Data":"3443129e3e0ade132d2f07767b542038259f11fa4d00f1cd2a206c9190d85c67"} Oct 01 16:15:43 crc kubenswrapper[4726]: I1001 16:15:43.163818 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.9120299589999998 podStartE2EDuration="5.163793112s" podCreationTimestamp="2025-10-01 16:15:38 +0000 UTC" firstStartedPulling="2025-10-01 16:15:39.137956279 +0000 UTC m=+992.039508866" lastFinishedPulling="2025-10-01 16:15:42.389719442 +0000 UTC m=+995.291272019" observedRunningTime="2025-10-01 16:15:43.152926276 +0000 UTC m=+996.054478863" watchObservedRunningTime="2025-10-01 16:15:43.163793112 +0000 UTC m=+996.065345689" Oct 01 16:15:43 crc kubenswrapper[4726]: I1001 16:15:43.429696 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-gj94c"] Oct 01 16:15:43 crc kubenswrapper[4726]: W1001 16:15:43.436798 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde31038e_fe8f_4d23_b875_aad3c596fbea.slice/crio-020ab1acaf9093f9aede40ed429ff3bc19e4bccebe020a65704e12339125d260 WatchSource:0}: Error finding container 020ab1acaf9093f9aede40ed429ff3bc19e4bccebe020a65704e12339125d260: Status 404 returned error can't find the container with id 020ab1acaf9093f9aede40ed429ff3bc19e4bccebe020a65704e12339125d260 Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.156036 4726 generic.go:334] "Generic (PLEG): container finished" podID="de31038e-fe8f-4d23-b875-aad3c596fbea" containerID="913c7a097b09c995bef106caf5cad831590479a065d7455706a5989f0e437bba" exitCode=0 Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.156222 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-gj94c" event={"ID":"de31038e-fe8f-4d23-b875-aad3c596fbea","Type":"ContainerDied","Data":"913c7a097b09c995bef106caf5cad831590479a065d7455706a5989f0e437bba"} Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.156519 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-gj94c" event={"ID":"de31038e-fe8f-4d23-b875-aad3c596fbea","Type":"ContainerStarted","Data":"020ab1acaf9093f9aede40ed429ff3bc19e4bccebe020a65704e12339125d260"} Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.477846 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.628481 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-dispersionconf\") pod \"4ab0d371-af59-407e-b596-3c533ad031b8\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.628526 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-combined-ca-bundle\") pod \"4ab0d371-af59-407e-b596-3c533ad031b8\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.628594 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4ab0d371-af59-407e-b596-3c533ad031b8-etc-swift\") pod \"4ab0d371-af59-407e-b596-3c533ad031b8\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.628631 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ab0d371-af59-407e-b596-3c533ad031b8-scripts\") pod \"4ab0d371-af59-407e-b596-3c533ad031b8\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.628665 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-swiftconf\") pod \"4ab0d371-af59-407e-b596-3c533ad031b8\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.628737 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4ab0d371-af59-407e-b596-3c533ad031b8-ring-data-devices\") pod \"4ab0d371-af59-407e-b596-3c533ad031b8\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.628777 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5ld6\" (UniqueName: \"kubernetes.io/projected/4ab0d371-af59-407e-b596-3c533ad031b8-kube-api-access-x5ld6\") pod \"4ab0d371-af59-407e-b596-3c533ad031b8\" (UID: \"4ab0d371-af59-407e-b596-3c533ad031b8\") " Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.629750 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ab0d371-af59-407e-b596-3c533ad031b8-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "4ab0d371-af59-407e-b596-3c533ad031b8" (UID: "4ab0d371-af59-407e-b596-3c533ad031b8"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.631186 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ab0d371-af59-407e-b596-3c533ad031b8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "4ab0d371-af59-407e-b596-3c533ad031b8" (UID: "4ab0d371-af59-407e-b596-3c533ad031b8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.633340 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ab0d371-af59-407e-b596-3c533ad031b8-kube-api-access-x5ld6" (OuterVolumeSpecName: "kube-api-access-x5ld6") pod "4ab0d371-af59-407e-b596-3c533ad031b8" (UID: "4ab0d371-af59-407e-b596-3c533ad031b8"). InnerVolumeSpecName "kube-api-access-x5ld6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.635829 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "4ab0d371-af59-407e-b596-3c533ad031b8" (UID: "4ab0d371-af59-407e-b596-3c533ad031b8"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.651179 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ab0d371-af59-407e-b596-3c533ad031b8" (UID: "4ab0d371-af59-407e-b596-3c533ad031b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.652270 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "4ab0d371-af59-407e-b596-3c533ad031b8" (UID: "4ab0d371-af59-407e-b596-3c533ad031b8"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.657111 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ab0d371-af59-407e-b596-3c533ad031b8-scripts" (OuterVolumeSpecName: "scripts") pod "4ab0d371-af59-407e-b596-3c533ad031b8" (UID: "4ab0d371-af59-407e-b596-3c533ad031b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.730922 4726 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.730956 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.730970 4726 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4ab0d371-af59-407e-b596-3c533ad031b8-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.730982 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ab0d371-af59-407e-b596-3c533ad031b8-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.730993 4726 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4ab0d371-af59-407e-b596-3c533ad031b8-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.731004 4726 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4ab0d371-af59-407e-b596-3c533ad031b8-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.731016 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5ld6\" (UniqueName: \"kubernetes.io/projected/4ab0d371-af59-407e-b596-3c533ad031b8-kube-api-access-x5ld6\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.791539 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 01 16:15:44 crc kubenswrapper[4726]: I1001 16:15:44.837312 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.131503 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-z6kcl"] Oct 01 16:15:45 crc kubenswrapper[4726]: E1001 16:15:45.132097 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab0d371-af59-407e-b596-3c533ad031b8" containerName="swift-ring-rebalance" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.132113 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab0d371-af59-407e-b596-3c533ad031b8" containerName="swift-ring-rebalance" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.132310 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab0d371-af59-407e-b596-3c533ad031b8" containerName="swift-ring-rebalance" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.132871 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-z6kcl" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.147595 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-z6kcl"] Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.165622 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t4hqx" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.165595 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t4hqx" event={"ID":"4ab0d371-af59-407e-b596-3c533ad031b8","Type":"ContainerDied","Data":"be9528e08f651009d87981c904062ede63770d83b92a4f9e3b3175e6a861df3b"} Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.165670 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be9528e08f651009d87981c904062ede63770d83b92a4f9e3b3175e6a861df3b" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.239523 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8bzv\" (UniqueName: \"kubernetes.io/projected/679126d8-15ab-4d02-bafe-72f753a5d526-kube-api-access-h8bzv\") pod \"barbican-db-create-z6kcl\" (UID: \"679126d8-15ab-4d02-bafe-72f753a5d526\") " pod="openstack/barbican-db-create-z6kcl" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.244686 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-m6p6g"] Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.246735 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-m6p6g" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.255327 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-m6p6g"] Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.341172 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8bzv\" (UniqueName: \"kubernetes.io/projected/679126d8-15ab-4d02-bafe-72f753a5d526-kube-api-access-h8bzv\") pod \"barbican-db-create-z6kcl\" (UID: \"679126d8-15ab-4d02-bafe-72f753a5d526\") " pod="openstack/barbican-db-create-z6kcl" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.341235 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2jtg\" (UniqueName: \"kubernetes.io/projected/5ca84aaa-d132-4716-9ec0-a6ccbf812ade-kube-api-access-c2jtg\") pod \"cinder-db-create-m6p6g\" (UID: \"5ca84aaa-d132-4716-9ec0-a6ccbf812ade\") " pod="openstack/cinder-db-create-m6p6g" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.372026 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8bzv\" (UniqueName: \"kubernetes.io/projected/679126d8-15ab-4d02-bafe-72f753a5d526-kube-api-access-h8bzv\") pod \"barbican-db-create-z6kcl\" (UID: \"679126d8-15ab-4d02-bafe-72f753a5d526\") " pod="openstack/barbican-db-create-z6kcl" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.425108 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-k9qcv"] Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.426125 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-k9qcv" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.438594 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-k9qcv"] Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.442316 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2jtg\" (UniqueName: \"kubernetes.io/projected/5ca84aaa-d132-4716-9ec0-a6ccbf812ade-kube-api-access-c2jtg\") pod \"cinder-db-create-m6p6g\" (UID: \"5ca84aaa-d132-4716-9ec0-a6ccbf812ade\") " pod="openstack/cinder-db-create-m6p6g" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.458675 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-z6kcl" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.471646 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2jtg\" (UniqueName: \"kubernetes.io/projected/5ca84aaa-d132-4716-9ec0-a6ccbf812ade-kube-api-access-c2jtg\") pod \"cinder-db-create-m6p6g\" (UID: \"5ca84aaa-d132-4716-9ec0-a6ccbf812ade\") " pod="openstack/cinder-db-create-m6p6g" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.543699 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t6j9\" (UniqueName: \"kubernetes.io/projected/ffa2e20c-77c7-436d-a548-f06977e9ae3e-kube-api-access-7t6j9\") pod \"neutron-db-create-k9qcv\" (UID: \"ffa2e20c-77c7-436d-a548-f06977e9ae3e\") " pod="openstack/neutron-db-create-k9qcv" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.551151 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gj94c" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.567541 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-m6p6g" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.645428 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58knx\" (UniqueName: \"kubernetes.io/projected/de31038e-fe8f-4d23-b875-aad3c596fbea-kube-api-access-58knx\") pod \"de31038e-fe8f-4d23-b875-aad3c596fbea\" (UID: \"de31038e-fe8f-4d23-b875-aad3c596fbea\") " Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.645868 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t6j9\" (UniqueName: \"kubernetes.io/projected/ffa2e20c-77c7-436d-a548-f06977e9ae3e-kube-api-access-7t6j9\") pod \"neutron-db-create-k9qcv\" (UID: \"ffa2e20c-77c7-436d-a548-f06977e9ae3e\") " pod="openstack/neutron-db-create-k9qcv" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.653134 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de31038e-fe8f-4d23-b875-aad3c596fbea-kube-api-access-58knx" (OuterVolumeSpecName: "kube-api-access-58knx") pod "de31038e-fe8f-4d23-b875-aad3c596fbea" (UID: "de31038e-fe8f-4d23-b875-aad3c596fbea"). InnerVolumeSpecName "kube-api-access-58knx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.676437 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t6j9\" (UniqueName: \"kubernetes.io/projected/ffa2e20c-77c7-436d-a548-f06977e9ae3e-kube-api-access-7t6j9\") pod \"neutron-db-create-k9qcv\" (UID: \"ffa2e20c-77c7-436d-a548-f06977e9ae3e\") " pod="openstack/neutron-db-create-k9qcv" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.751996 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58knx\" (UniqueName: \"kubernetes.io/projected/de31038e-fe8f-4d23-b875-aad3c596fbea-kube-api-access-58knx\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.752132 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-k9qcv" Oct 01 16:15:45 crc kubenswrapper[4726]: I1001 16:15:45.945866 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-z6kcl"] Oct 01 16:15:45 crc kubenswrapper[4726]: W1001 16:15:45.953404 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod679126d8_15ab_4d02_bafe_72f753a5d526.slice/crio-c0fcfdd6e7e0dcb10ba4cbf6ec0996cc9c2ae943de67db0f571ad93d4499d30b WatchSource:0}: Error finding container c0fcfdd6e7e0dcb10ba4cbf6ec0996cc9c2ae943de67db0f571ad93d4499d30b: Status 404 returned error can't find the container with id c0fcfdd6e7e0dcb10ba4cbf6ec0996cc9c2ae943de67db0f571ad93d4499d30b Oct 01 16:15:46 crc kubenswrapper[4726]: I1001 16:15:46.046920 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-m6p6g"] Oct 01 16:15:46 crc kubenswrapper[4726]: W1001 16:15:46.048657 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ca84aaa_d132_4716_9ec0_a6ccbf812ade.slice/crio-83e091b92fb671319df4a3669d7d98e7fb537a7f86f3b912779059a1d8bd6c62 WatchSource:0}: Error finding container 83e091b92fb671319df4a3669d7d98e7fb537a7f86f3b912779059a1d8bd6c62: Status 404 returned error can't find the container with id 83e091b92fb671319df4a3669d7d98e7fb537a7f86f3b912779059a1d8bd6c62 Oct 01 16:15:46 crc kubenswrapper[4726]: I1001 16:15:46.176264 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-m6p6g" event={"ID":"5ca84aaa-d132-4716-9ec0-a6ccbf812ade","Type":"ContainerStarted","Data":"83e091b92fb671319df4a3669d7d98e7fb537a7f86f3b912779059a1d8bd6c62"} Oct 01 16:15:46 crc kubenswrapper[4726]: I1001 16:15:46.178123 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-z6kcl" event={"ID":"679126d8-15ab-4d02-bafe-72f753a5d526","Type":"ContainerStarted","Data":"c0fcfdd6e7e0dcb10ba4cbf6ec0996cc9c2ae943de67db0f571ad93d4499d30b"} Oct 01 16:15:46 crc kubenswrapper[4726]: I1001 16:15:46.180847 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-gj94c" event={"ID":"de31038e-fe8f-4d23-b875-aad3c596fbea","Type":"ContainerDied","Data":"020ab1acaf9093f9aede40ed429ff3bc19e4bccebe020a65704e12339125d260"} Oct 01 16:15:46 crc kubenswrapper[4726]: I1001 16:15:46.180882 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="020ab1acaf9093f9aede40ed429ff3bc19e4bccebe020a65704e12339125d260" Oct 01 16:15:46 crc kubenswrapper[4726]: I1001 16:15:46.180947 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gj94c" Oct 01 16:15:46 crc kubenswrapper[4726]: I1001 16:15:46.183833 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-k9qcv"] Oct 01 16:15:46 crc kubenswrapper[4726]: W1001 16:15:46.195327 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffa2e20c_77c7_436d_a548_f06977e9ae3e.slice/crio-91af32b1853d257a760b31f6b4e5abd0bdcfa5c5cd36a627ae98c1297767a61a WatchSource:0}: Error finding container 91af32b1853d257a760b31f6b4e5abd0bdcfa5c5cd36a627ae98c1297767a61a: Status 404 returned error can't find the container with id 91af32b1853d257a760b31f6b4e5abd0bdcfa5c5cd36a627ae98c1297767a61a Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.224974 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-9457-account-create-6wcqq"] Oct 01 16:15:47 crc kubenswrapper[4726]: E1001 16:15:47.225579 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de31038e-fe8f-4d23-b875-aad3c596fbea" containerName="mariadb-database-create" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.225590 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="de31038e-fe8f-4d23-b875-aad3c596fbea" containerName="mariadb-database-create" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.225744 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="de31038e-fe8f-4d23-b875-aad3c596fbea" containerName="mariadb-database-create" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.226854 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9457-account-create-6wcqq" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.232313 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.233622 4726 generic.go:334] "Generic (PLEG): container finished" podID="ffa2e20c-77c7-436d-a548-f06977e9ae3e" containerID="00ef31a6c2ac6402d402e0304aad181a6860208742aeb4be38bb86509da08d98" exitCode=0 Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.233709 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-k9qcv" event={"ID":"ffa2e20c-77c7-436d-a548-f06977e9ae3e","Type":"ContainerDied","Data":"00ef31a6c2ac6402d402e0304aad181a6860208742aeb4be38bb86509da08d98"} Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.233861 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-k9qcv" event={"ID":"ffa2e20c-77c7-436d-a548-f06977e9ae3e","Type":"ContainerStarted","Data":"91af32b1853d257a760b31f6b4e5abd0bdcfa5c5cd36a627ae98c1297767a61a"} Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.235466 4726 generic.go:334] "Generic (PLEG): container finished" podID="679126d8-15ab-4d02-bafe-72f753a5d526" containerID="a4eed3fca647d776718f3159f4644836840c3c9cee56d8b1260375375127d5e5" exitCode=0 Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.235573 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-z6kcl" event={"ID":"679126d8-15ab-4d02-bafe-72f753a5d526","Type":"ContainerDied","Data":"a4eed3fca647d776718f3159f4644836840c3c9cee56d8b1260375375127d5e5"} Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.236935 4726 generic.go:334] "Generic (PLEG): container finished" podID="5ca84aaa-d132-4716-9ec0-a6ccbf812ade" containerID="61b7397e92066dbc8c6f9e5041df59102d2d6a4422ec5a955752453831739c9e" exitCode=0 Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.236967 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-m6p6g" event={"ID":"5ca84aaa-d132-4716-9ec0-a6ccbf812ade","Type":"ContainerDied","Data":"61b7397e92066dbc8c6f9e5041df59102d2d6a4422ec5a955752453831739c9e"} Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.243910 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9457-account-create-6wcqq"] Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.384923 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-268tv\" (UniqueName: \"kubernetes.io/projected/6b9a5008-3aed-4f59-9ad5-3f568f35c0dc-kube-api-access-268tv\") pod \"keystone-9457-account-create-6wcqq\" (UID: \"6b9a5008-3aed-4f59-9ad5-3f568f35c0dc\") " pod="openstack/keystone-9457-account-create-6wcqq" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.486271 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-268tv\" (UniqueName: \"kubernetes.io/projected/6b9a5008-3aed-4f59-9ad5-3f568f35c0dc-kube-api-access-268tv\") pod \"keystone-9457-account-create-6wcqq\" (UID: \"6b9a5008-3aed-4f59-9ad5-3f568f35c0dc\") " pod="openstack/keystone-9457-account-create-6wcqq" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.515961 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-268tv\" (UniqueName: \"kubernetes.io/projected/6b9a5008-3aed-4f59-9ad5-3f568f35c0dc-kube-api-access-268tv\") pod \"keystone-9457-account-create-6wcqq\" (UID: \"6b9a5008-3aed-4f59-9ad5-3f568f35c0dc\") " pod="openstack/keystone-9457-account-create-6wcqq" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.530501 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9bb7-account-create-xzkf9"] Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.531448 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9bb7-account-create-xzkf9" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.533826 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.549223 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9bb7-account-create-xzkf9"] Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.596141 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9457-account-create-6wcqq" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.690014 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hvwx\" (UniqueName: \"kubernetes.io/projected/68b428fe-6f95-4fbd-b31a-b3aafba90d32-kube-api-access-8hvwx\") pod \"placement-9bb7-account-create-xzkf9\" (UID: \"68b428fe-6f95-4fbd-b31a-b3aafba90d32\") " pod="openstack/placement-9bb7-account-create-xzkf9" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.791366 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hvwx\" (UniqueName: \"kubernetes.io/projected/68b428fe-6f95-4fbd-b31a-b3aafba90d32-kube-api-access-8hvwx\") pod \"placement-9bb7-account-create-xzkf9\" (UID: \"68b428fe-6f95-4fbd-b31a-b3aafba90d32\") " pod="openstack/placement-9bb7-account-create-xzkf9" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.812035 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hvwx\" (UniqueName: \"kubernetes.io/projected/68b428fe-6f95-4fbd-b31a-b3aafba90d32-kube-api-access-8hvwx\") pod \"placement-9bb7-account-create-xzkf9\" (UID: \"68b428fe-6f95-4fbd-b31a-b3aafba90d32\") " pod="openstack/placement-9bb7-account-create-xzkf9" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.867104 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9bb7-account-create-xzkf9" Oct 01 16:15:47 crc kubenswrapper[4726]: I1001 16:15:47.867762 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-z4lk4" podUID="f67ec6c1-adf7-47f9-855f-fde38e0430c7" containerName="ovn-controller" probeResult="failure" output=< Oct 01 16:15:47 crc kubenswrapper[4726]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 01 16:15:47 crc kubenswrapper[4726]: > Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.047345 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9457-account-create-6wcqq"] Oct 01 16:15:48 crc kubenswrapper[4726]: W1001 16:15:48.056474 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b9a5008_3aed_4f59_9ad5_3f568f35c0dc.slice/crio-abba43adfa7f48d4f13c168020b0cb4b1ccf41225dbaceca31a860a39a607950 WatchSource:0}: Error finding container abba43adfa7f48d4f13c168020b0cb4b1ccf41225dbaceca31a860a39a607950: Status 404 returned error can't find the container with id abba43adfa7f48d4f13c168020b0cb4b1ccf41225dbaceca31a860a39a607950 Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.245604 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9457-account-create-6wcqq" event={"ID":"6b9a5008-3aed-4f59-9ad5-3f568f35c0dc","Type":"ContainerStarted","Data":"5acce15b79fcc32e2b7f5ff77c8e1e597a20080a3c4105db28970445022b8206"} Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.245653 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9457-account-create-6wcqq" event={"ID":"6b9a5008-3aed-4f59-9ad5-3f568f35c0dc","Type":"ContainerStarted","Data":"abba43adfa7f48d4f13c168020b0cb4b1ccf41225dbaceca31a860a39a607950"} Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.265755 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-9457-account-create-6wcqq" podStartSLOduration=1.265734015 podStartE2EDuration="1.265734015s" podCreationTimestamp="2025-10-01 16:15:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:15:48.261502552 +0000 UTC m=+1001.163055139" watchObservedRunningTime="2025-10-01 16:15:48.265734015 +0000 UTC m=+1001.167286592" Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.316768 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9bb7-account-create-xzkf9"] Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.669604 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-z6kcl" Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.809415 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8bzv\" (UniqueName: \"kubernetes.io/projected/679126d8-15ab-4d02-bafe-72f753a5d526-kube-api-access-h8bzv\") pod \"679126d8-15ab-4d02-bafe-72f753a5d526\" (UID: \"679126d8-15ab-4d02-bafe-72f753a5d526\") " Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.814282 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/679126d8-15ab-4d02-bafe-72f753a5d526-kube-api-access-h8bzv" (OuterVolumeSpecName: "kube-api-access-h8bzv") pod "679126d8-15ab-4d02-bafe-72f753a5d526" (UID: "679126d8-15ab-4d02-bafe-72f753a5d526"). InnerVolumeSpecName "kube-api-access-h8bzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.854136 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-m6p6g" Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.862469 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-k9qcv" Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.910575 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2jtg\" (UniqueName: \"kubernetes.io/projected/5ca84aaa-d132-4716-9ec0-a6ccbf812ade-kube-api-access-c2jtg\") pod \"5ca84aaa-d132-4716-9ec0-a6ccbf812ade\" (UID: \"5ca84aaa-d132-4716-9ec0-a6ccbf812ade\") " Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.910630 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t6j9\" (UniqueName: \"kubernetes.io/projected/ffa2e20c-77c7-436d-a548-f06977e9ae3e-kube-api-access-7t6j9\") pod \"ffa2e20c-77c7-436d-a548-f06977e9ae3e\" (UID: \"ffa2e20c-77c7-436d-a548-f06977e9ae3e\") " Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.911091 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8bzv\" (UniqueName: \"kubernetes.io/projected/679126d8-15ab-4d02-bafe-72f753a5d526-kube-api-access-h8bzv\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.913883 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffa2e20c-77c7-436d-a548-f06977e9ae3e-kube-api-access-7t6j9" (OuterVolumeSpecName: "kube-api-access-7t6j9") pod "ffa2e20c-77c7-436d-a548-f06977e9ae3e" (UID: "ffa2e20c-77c7-436d-a548-f06977e9ae3e"). InnerVolumeSpecName "kube-api-access-7t6j9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:48 crc kubenswrapper[4726]: I1001 16:15:48.913920 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ca84aaa-d132-4716-9ec0-a6ccbf812ade-kube-api-access-c2jtg" (OuterVolumeSpecName: "kube-api-access-c2jtg") pod "5ca84aaa-d132-4716-9ec0-a6ccbf812ade" (UID: "5ca84aaa-d132-4716-9ec0-a6ccbf812ade"). InnerVolumeSpecName "kube-api-access-c2jtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.012835 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2jtg\" (UniqueName: \"kubernetes.io/projected/5ca84aaa-d132-4716-9ec0-a6ccbf812ade-kube-api-access-c2jtg\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.012869 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t6j9\" (UniqueName: \"kubernetes.io/projected/ffa2e20c-77c7-436d-a548-f06977e9ae3e-kube-api-access-7t6j9\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.254876 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-k9qcv" event={"ID":"ffa2e20c-77c7-436d-a548-f06977e9ae3e","Type":"ContainerDied","Data":"91af32b1853d257a760b31f6b4e5abd0bdcfa5c5cd36a627ae98c1297767a61a"} Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.254913 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91af32b1853d257a760b31f6b4e5abd0bdcfa5c5cd36a627ae98c1297767a61a" Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.254925 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-k9qcv" Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.256246 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-z6kcl" Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.256243 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-z6kcl" event={"ID":"679126d8-15ab-4d02-bafe-72f753a5d526","Type":"ContainerDied","Data":"c0fcfdd6e7e0dcb10ba4cbf6ec0996cc9c2ae943de67db0f571ad93d4499d30b"} Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.256357 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0fcfdd6e7e0dcb10ba4cbf6ec0996cc9c2ae943de67db0f571ad93d4499d30b" Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.257717 4726 generic.go:334] "Generic (PLEG): container finished" podID="68b428fe-6f95-4fbd-b31a-b3aafba90d32" containerID="606f5e42642c8a16c99a89877d19df62fa1d606137bd5c7bd896ad74f51bbf27" exitCode=0 Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.258145 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9bb7-account-create-xzkf9" event={"ID":"68b428fe-6f95-4fbd-b31a-b3aafba90d32","Type":"ContainerDied","Data":"606f5e42642c8a16c99a89877d19df62fa1d606137bd5c7bd896ad74f51bbf27"} Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.258179 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9bb7-account-create-xzkf9" event={"ID":"68b428fe-6f95-4fbd-b31a-b3aafba90d32","Type":"ContainerStarted","Data":"34d1278b0aa47fb966026953a77bcdc8f48d069a2350904e4217e707b7fe0859"} Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.259518 4726 generic.go:334] "Generic (PLEG): container finished" podID="6b9a5008-3aed-4f59-9ad5-3f568f35c0dc" containerID="5acce15b79fcc32e2b7f5ff77c8e1e597a20080a3c4105db28970445022b8206" exitCode=0 Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.259599 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9457-account-create-6wcqq" event={"ID":"6b9a5008-3aed-4f59-9ad5-3f568f35c0dc","Type":"ContainerDied","Data":"5acce15b79fcc32e2b7f5ff77c8e1e597a20080a3c4105db28970445022b8206"} Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.262559 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-m6p6g" event={"ID":"5ca84aaa-d132-4716-9ec0-a6ccbf812ade","Type":"ContainerDied","Data":"83e091b92fb671319df4a3669d7d98e7fb537a7f86f3b912779059a1d8bd6c62"} Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.262594 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83e091b92fb671319df4a3669d7d98e7fb537a7f86f3b912779059a1d8bd6c62" Oct 01 16:15:49 crc kubenswrapper[4726]: I1001 16:15:49.262645 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-m6p6g" Oct 01 16:15:50 crc kubenswrapper[4726]: I1001 16:15:50.708393 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9457-account-create-6wcqq" Oct 01 16:15:50 crc kubenswrapper[4726]: I1001 16:15:50.713939 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9bb7-account-create-xzkf9" Oct 01 16:15:50 crc kubenswrapper[4726]: I1001 16:15:50.843538 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hvwx\" (UniqueName: \"kubernetes.io/projected/68b428fe-6f95-4fbd-b31a-b3aafba90d32-kube-api-access-8hvwx\") pod \"68b428fe-6f95-4fbd-b31a-b3aafba90d32\" (UID: \"68b428fe-6f95-4fbd-b31a-b3aafba90d32\") " Oct 01 16:15:50 crc kubenswrapper[4726]: I1001 16:15:50.843826 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-268tv\" (UniqueName: \"kubernetes.io/projected/6b9a5008-3aed-4f59-9ad5-3f568f35c0dc-kube-api-access-268tv\") pod \"6b9a5008-3aed-4f59-9ad5-3f568f35c0dc\" (UID: \"6b9a5008-3aed-4f59-9ad5-3f568f35c0dc\") " Oct 01 16:15:50 crc kubenswrapper[4726]: I1001 16:15:50.850206 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68b428fe-6f95-4fbd-b31a-b3aafba90d32-kube-api-access-8hvwx" (OuterVolumeSpecName: "kube-api-access-8hvwx") pod "68b428fe-6f95-4fbd-b31a-b3aafba90d32" (UID: "68b428fe-6f95-4fbd-b31a-b3aafba90d32"). InnerVolumeSpecName "kube-api-access-8hvwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:50 crc kubenswrapper[4726]: I1001 16:15:50.850289 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b9a5008-3aed-4f59-9ad5-3f568f35c0dc-kube-api-access-268tv" (OuterVolumeSpecName: "kube-api-access-268tv") pod "6b9a5008-3aed-4f59-9ad5-3f568f35c0dc" (UID: "6b9a5008-3aed-4f59-9ad5-3f568f35c0dc"). InnerVolumeSpecName "kube-api-access-268tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:50 crc kubenswrapper[4726]: I1001 16:15:50.946067 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-268tv\" (UniqueName: \"kubernetes.io/projected/6b9a5008-3aed-4f59-9ad5-3f568f35c0dc-kube-api-access-268tv\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:50 crc kubenswrapper[4726]: I1001 16:15:50.946108 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hvwx\" (UniqueName: \"kubernetes.io/projected/68b428fe-6f95-4fbd-b31a-b3aafba90d32-kube-api-access-8hvwx\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:51 crc kubenswrapper[4726]: I1001 16:15:51.283215 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9bb7-account-create-xzkf9" Oct 01 16:15:51 crc kubenswrapper[4726]: I1001 16:15:51.283207 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9bb7-account-create-xzkf9" event={"ID":"68b428fe-6f95-4fbd-b31a-b3aafba90d32","Type":"ContainerDied","Data":"34d1278b0aa47fb966026953a77bcdc8f48d069a2350904e4217e707b7fe0859"} Oct 01 16:15:51 crc kubenswrapper[4726]: I1001 16:15:51.284215 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34d1278b0aa47fb966026953a77bcdc8f48d069a2350904e4217e707b7fe0859" Oct 01 16:15:51 crc kubenswrapper[4726]: I1001 16:15:51.285082 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9457-account-create-6wcqq" event={"ID":"6b9a5008-3aed-4f59-9ad5-3f568f35c0dc","Type":"ContainerDied","Data":"abba43adfa7f48d4f13c168020b0cb4b1ccf41225dbaceca31a860a39a607950"} Oct 01 16:15:51 crc kubenswrapper[4726]: I1001 16:15:51.285127 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abba43adfa7f48d4f13c168020b0cb4b1ccf41225dbaceca31a860a39a607950" Oct 01 16:15:51 crc kubenswrapper[4726]: I1001 16:15:51.285222 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9457-account-create-6wcqq" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.751561 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-6f33-account-create-rw9w7"] Oct 01 16:15:52 crc kubenswrapper[4726]: E1001 16:15:52.752124 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ca84aaa-d132-4716-9ec0-a6ccbf812ade" containerName="mariadb-database-create" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.752137 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ca84aaa-d132-4716-9ec0-a6ccbf812ade" containerName="mariadb-database-create" Oct 01 16:15:52 crc kubenswrapper[4726]: E1001 16:15:52.752153 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b9a5008-3aed-4f59-9ad5-3f568f35c0dc" containerName="mariadb-account-create" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.752158 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b9a5008-3aed-4f59-9ad5-3f568f35c0dc" containerName="mariadb-account-create" Oct 01 16:15:52 crc kubenswrapper[4726]: E1001 16:15:52.752173 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="679126d8-15ab-4d02-bafe-72f753a5d526" containerName="mariadb-database-create" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.752180 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="679126d8-15ab-4d02-bafe-72f753a5d526" containerName="mariadb-database-create" Oct 01 16:15:52 crc kubenswrapper[4726]: E1001 16:15:52.752203 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b428fe-6f95-4fbd-b31a-b3aafba90d32" containerName="mariadb-account-create" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.752209 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b428fe-6f95-4fbd-b31a-b3aafba90d32" containerName="mariadb-account-create" Oct 01 16:15:52 crc kubenswrapper[4726]: E1001 16:15:52.752217 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa2e20c-77c7-436d-a548-f06977e9ae3e" containerName="mariadb-database-create" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.752223 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa2e20c-77c7-436d-a548-f06977e9ae3e" containerName="mariadb-database-create" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.752371 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ca84aaa-d132-4716-9ec0-a6ccbf812ade" containerName="mariadb-database-create" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.752386 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b9a5008-3aed-4f59-9ad5-3f568f35c0dc" containerName="mariadb-account-create" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.752394 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b428fe-6f95-4fbd-b31a-b3aafba90d32" containerName="mariadb-account-create" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.752411 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="679126d8-15ab-4d02-bafe-72f753a5d526" containerName="mariadb-database-create" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.752419 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa2e20c-77c7-436d-a548-f06977e9ae3e" containerName="mariadb-database-create" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.752910 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6f33-account-create-rw9w7" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.755261 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.771931 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6f33-account-create-rw9w7"] Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.848522 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-2df5f"] Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.849763 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-2df5f" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.862208 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.862208 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5q5v6" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.862323 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.862334 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.879022 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpknm\" (UniqueName: \"kubernetes.io/projected/50c724bf-9bf3-4e8b-97c5-09199a794376-kube-api-access-kpknm\") pod \"glance-6f33-account-create-rw9w7\" (UID: \"50c724bf-9bf3-4e8b-97c5-09199a794376\") " pod="openstack/glance-6f33-account-create-rw9w7" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.887117 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-2df5f"] Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.981164 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87hgb\" (UniqueName: \"kubernetes.io/projected/4a875fd8-855b-48b4-91ad-043ad70711a1-kube-api-access-87hgb\") pod \"keystone-db-sync-2df5f\" (UID: \"4a875fd8-855b-48b4-91ad-043ad70711a1\") " pod="openstack/keystone-db-sync-2df5f" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.981286 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a875fd8-855b-48b4-91ad-043ad70711a1-combined-ca-bundle\") pod \"keystone-db-sync-2df5f\" (UID: \"4a875fd8-855b-48b4-91ad-043ad70711a1\") " pod="openstack/keystone-db-sync-2df5f" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.981370 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpknm\" (UniqueName: \"kubernetes.io/projected/50c724bf-9bf3-4e8b-97c5-09199a794376-kube-api-access-kpknm\") pod \"glance-6f33-account-create-rw9w7\" (UID: \"50c724bf-9bf3-4e8b-97c5-09199a794376\") " pod="openstack/glance-6f33-account-create-rw9w7" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.981405 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a875fd8-855b-48b4-91ad-043ad70711a1-config-data\") pod \"keystone-db-sync-2df5f\" (UID: \"4a875fd8-855b-48b4-91ad-043ad70711a1\") " pod="openstack/keystone-db-sync-2df5f" Oct 01 16:15:52 crc kubenswrapper[4726]: I1001 16:15:52.986634 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-z4lk4" podUID="f67ec6c1-adf7-47f9-855f-fde38e0430c7" containerName="ovn-controller" probeResult="failure" output=< Oct 01 16:15:52 crc kubenswrapper[4726]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 01 16:15:52 crc kubenswrapper[4726]: > Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.019715 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpknm\" (UniqueName: \"kubernetes.io/projected/50c724bf-9bf3-4e8b-97c5-09199a794376-kube-api-access-kpknm\") pod \"glance-6f33-account-create-rw9w7\" (UID: \"50c724bf-9bf3-4e8b-97c5-09199a794376\") " pod="openstack/glance-6f33-account-create-rw9w7" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.044159 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-4msxz" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.074314 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6f33-account-create-rw9w7" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.082987 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87hgb\" (UniqueName: \"kubernetes.io/projected/4a875fd8-855b-48b4-91ad-043ad70711a1-kube-api-access-87hgb\") pod \"keystone-db-sync-2df5f\" (UID: \"4a875fd8-855b-48b4-91ad-043ad70711a1\") " pod="openstack/keystone-db-sync-2df5f" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.083083 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a875fd8-855b-48b4-91ad-043ad70711a1-combined-ca-bundle\") pod \"keystone-db-sync-2df5f\" (UID: \"4a875fd8-855b-48b4-91ad-043ad70711a1\") " pod="openstack/keystone-db-sync-2df5f" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.083138 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a875fd8-855b-48b4-91ad-043ad70711a1-config-data\") pod \"keystone-db-sync-2df5f\" (UID: \"4a875fd8-855b-48b4-91ad-043ad70711a1\") " pod="openstack/keystone-db-sync-2df5f" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.087738 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a875fd8-855b-48b4-91ad-043ad70711a1-combined-ca-bundle\") pod \"keystone-db-sync-2df5f\" (UID: \"4a875fd8-855b-48b4-91ad-043ad70711a1\") " pod="openstack/keystone-db-sync-2df5f" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.088553 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a875fd8-855b-48b4-91ad-043ad70711a1-config-data\") pod \"keystone-db-sync-2df5f\" (UID: \"4a875fd8-855b-48b4-91ad-043ad70711a1\") " pod="openstack/keystone-db-sync-2df5f" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.104636 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87hgb\" (UniqueName: \"kubernetes.io/projected/4a875fd8-855b-48b4-91ad-043ad70711a1-kube-api-access-87hgb\") pod \"keystone-db-sync-2df5f\" (UID: \"4a875fd8-855b-48b4-91ad-043ad70711a1\") " pod="openstack/keystone-db-sync-2df5f" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.181187 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-2df5f" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.250661 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-z4lk4-config-qmmd8"] Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.252075 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.257802 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.263510 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z4lk4-config-qmmd8"] Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.387728 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/602d1e20-fa25-4183-9801-abadbbbe3495-scripts\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.388153 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8kpg\" (UniqueName: \"kubernetes.io/projected/602d1e20-fa25-4183-9801-abadbbbe3495-kube-api-access-q8kpg\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.388232 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-run\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.388429 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-run-ovn\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.388484 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/602d1e20-fa25-4183-9801-abadbbbe3495-additional-scripts\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.388539 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-log-ovn\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.489522 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-run-ovn\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.489595 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/602d1e20-fa25-4183-9801-abadbbbe3495-additional-scripts\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.489631 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-log-ovn\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.489677 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/602d1e20-fa25-4183-9801-abadbbbe3495-scripts\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.489707 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8kpg\" (UniqueName: \"kubernetes.io/projected/602d1e20-fa25-4183-9801-abadbbbe3495-kube-api-access-q8kpg\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.489731 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-run\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.489780 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-log-ovn\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.489780 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-run-ovn\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.489855 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-run\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.490636 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/602d1e20-fa25-4183-9801-abadbbbe3495-additional-scripts\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.491779 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/602d1e20-fa25-4183-9801-abadbbbe3495-scripts\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.508971 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8kpg\" (UniqueName: \"kubernetes.io/projected/602d1e20-fa25-4183-9801-abadbbbe3495-kube-api-access-q8kpg\") pod \"ovn-controller-z4lk4-config-qmmd8\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.596924 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.603561 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6f33-account-create-rw9w7"] Oct 01 16:15:53 crc kubenswrapper[4726]: W1001 16:15:53.613725 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50c724bf_9bf3_4e8b_97c5_09199a794376.slice/crio-fe7ba6a0250898750a90261e74412b10d5968470b6ca2000c9aac04fcfeba241 WatchSource:0}: Error finding container fe7ba6a0250898750a90261e74412b10d5968470b6ca2000c9aac04fcfeba241: Status 404 returned error can't find the container with id fe7ba6a0250898750a90261e74412b10d5968470b6ca2000c9aac04fcfeba241 Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.665824 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-2df5f"] Oct 01 16:15:53 crc kubenswrapper[4726]: I1001 16:15:53.766694 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 01 16:15:54 crc kubenswrapper[4726]: I1001 16:15:54.025725 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z4lk4-config-qmmd8"] Oct 01 16:15:54 crc kubenswrapper[4726]: W1001 16:15:54.035462 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod602d1e20_fa25_4183_9801_abadbbbe3495.slice/crio-1a88cbef062fb21a5ef20e55ef413499e2b023c1f21493626823013d56c23cbc WatchSource:0}: Error finding container 1a88cbef062fb21a5ef20e55ef413499e2b023c1f21493626823013d56c23cbc: Status 404 returned error can't find the container with id 1a88cbef062fb21a5ef20e55ef413499e2b023c1f21493626823013d56c23cbc Oct 01 16:15:54 crc kubenswrapper[4726]: I1001 16:15:54.316418 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-2df5f" event={"ID":"4a875fd8-855b-48b4-91ad-043ad70711a1","Type":"ContainerStarted","Data":"33a6c1e135279a195ed32f2aab1ff9a5ae4c860960fd388fd910b5cba8cb8e52"} Oct 01 16:15:54 crc kubenswrapper[4726]: I1001 16:15:54.317620 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z4lk4-config-qmmd8" event={"ID":"602d1e20-fa25-4183-9801-abadbbbe3495","Type":"ContainerStarted","Data":"1a88cbef062fb21a5ef20e55ef413499e2b023c1f21493626823013d56c23cbc"} Oct 01 16:15:54 crc kubenswrapper[4726]: I1001 16:15:54.319965 4726 generic.go:334] "Generic (PLEG): container finished" podID="50c724bf-9bf3-4e8b-97c5-09199a794376" containerID="778ae1d844e80e8a01605ecdf5b8268efe10337575ffb4ec750b9f8d0e97c270" exitCode=0 Oct 01 16:15:54 crc kubenswrapper[4726]: I1001 16:15:54.319997 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6f33-account-create-rw9w7" event={"ID":"50c724bf-9bf3-4e8b-97c5-09199a794376","Type":"ContainerDied","Data":"778ae1d844e80e8a01605ecdf5b8268efe10337575ffb4ec750b9f8d0e97c270"} Oct 01 16:15:54 crc kubenswrapper[4726]: I1001 16:15:54.320016 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6f33-account-create-rw9w7" event={"ID":"50c724bf-9bf3-4e8b-97c5-09199a794376","Type":"ContainerStarted","Data":"fe7ba6a0250898750a90261e74412b10d5968470b6ca2000c9aac04fcfeba241"} Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.121136 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-7b0c-account-create-jf4fq"] Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.124130 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7b0c-account-create-jf4fq" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.125803 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.127984 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7b0c-account-create-jf4fq"] Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.220812 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qdbv\" (UniqueName: \"kubernetes.io/projected/56bde538-7fa3-4a9b-be9f-87ee018e6a43-kube-api-access-5qdbv\") pod \"barbican-7b0c-account-create-jf4fq\" (UID: \"56bde538-7fa3-4a9b-be9f-87ee018e6a43\") " pod="openstack/barbican-7b0c-account-create-jf4fq" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.317755 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-6bf9-account-create-9xm7t"] Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.319477 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6bf9-account-create-9xm7t" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.321457 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.322894 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qdbv\" (UniqueName: \"kubernetes.io/projected/56bde538-7fa3-4a9b-be9f-87ee018e6a43-kube-api-access-5qdbv\") pod \"barbican-7b0c-account-create-jf4fq\" (UID: \"56bde538-7fa3-4a9b-be9f-87ee018e6a43\") " pod="openstack/barbican-7b0c-account-create-jf4fq" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.330438 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6bf9-account-create-9xm7t"] Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.348418 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qdbv\" (UniqueName: \"kubernetes.io/projected/56bde538-7fa3-4a9b-be9f-87ee018e6a43-kube-api-access-5qdbv\") pod \"barbican-7b0c-account-create-jf4fq\" (UID: \"56bde538-7fa3-4a9b-be9f-87ee018e6a43\") " pod="openstack/barbican-7b0c-account-create-jf4fq" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.349858 4726 generic.go:334] "Generic (PLEG): container finished" podID="602d1e20-fa25-4183-9801-abadbbbe3495" containerID="2f52687b6d1be2679e117f20e531f467956a369117c02dd253f419d21137f23b" exitCode=0 Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.350394 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z4lk4-config-qmmd8" event={"ID":"602d1e20-fa25-4183-9801-abadbbbe3495","Type":"ContainerDied","Data":"2f52687b6d1be2679e117f20e531f467956a369117c02dd253f419d21137f23b"} Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.428147 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99424\" (UniqueName: \"kubernetes.io/projected/1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4-kube-api-access-99424\") pod \"cinder-6bf9-account-create-9xm7t\" (UID: \"1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4\") " pod="openstack/cinder-6bf9-account-create-9xm7t" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.449220 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7b0c-account-create-jf4fq" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.523490 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-1b50-account-create-qw2dx"] Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.525531 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-1b50-account-create-qw2dx" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.528196 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.529199 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99424\" (UniqueName: \"kubernetes.io/projected/1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4-kube-api-access-99424\") pod \"cinder-6bf9-account-create-9xm7t\" (UID: \"1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4\") " pod="openstack/cinder-6bf9-account-create-9xm7t" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.532003 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-1b50-account-create-qw2dx"] Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.553567 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99424\" (UniqueName: \"kubernetes.io/projected/1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4-kube-api-access-99424\") pod \"cinder-6bf9-account-create-9xm7t\" (UID: \"1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4\") " pod="openstack/cinder-6bf9-account-create-9xm7t" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.630862 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7jh8\" (UniqueName: \"kubernetes.io/projected/708ea401-1d87-4a51-ab33-b9318fe542e2-kube-api-access-w7jh8\") pod \"neutron-1b50-account-create-qw2dx\" (UID: \"708ea401-1d87-4a51-ab33-b9318fe542e2\") " pod="openstack/neutron-1b50-account-create-qw2dx" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.659204 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6f33-account-create-rw9w7" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.732244 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6bf9-account-create-9xm7t" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.733995 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpknm\" (UniqueName: \"kubernetes.io/projected/50c724bf-9bf3-4e8b-97c5-09199a794376-kube-api-access-kpknm\") pod \"50c724bf-9bf3-4e8b-97c5-09199a794376\" (UID: \"50c724bf-9bf3-4e8b-97c5-09199a794376\") " Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.734622 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7jh8\" (UniqueName: \"kubernetes.io/projected/708ea401-1d87-4a51-ab33-b9318fe542e2-kube-api-access-w7jh8\") pod \"neutron-1b50-account-create-qw2dx\" (UID: \"708ea401-1d87-4a51-ab33-b9318fe542e2\") " pod="openstack/neutron-1b50-account-create-qw2dx" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.737340 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50c724bf-9bf3-4e8b-97c5-09199a794376-kube-api-access-kpknm" (OuterVolumeSpecName: "kube-api-access-kpknm") pod "50c724bf-9bf3-4e8b-97c5-09199a794376" (UID: "50c724bf-9bf3-4e8b-97c5-09199a794376"). InnerVolumeSpecName "kube-api-access-kpknm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.755990 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7jh8\" (UniqueName: \"kubernetes.io/projected/708ea401-1d87-4a51-ab33-b9318fe542e2-kube-api-access-w7jh8\") pod \"neutron-1b50-account-create-qw2dx\" (UID: \"708ea401-1d87-4a51-ab33-b9318fe542e2\") " pod="openstack/neutron-1b50-account-create-qw2dx" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.835925 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpknm\" (UniqueName: \"kubernetes.io/projected/50c724bf-9bf3-4e8b-97c5-09199a794376-kube-api-access-kpknm\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.945957 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7b0c-account-create-jf4fq"] Oct 01 16:15:55 crc kubenswrapper[4726]: W1001 16:15:55.954243 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56bde538_7fa3_4a9b_be9f_87ee018e6a43.slice/crio-4c08efe1a2fb3f7a54cfcfbb81890fb6139c1569d2dc8826629dc5d40bd3baf6 WatchSource:0}: Error finding container 4c08efe1a2fb3f7a54cfcfbb81890fb6139c1569d2dc8826629dc5d40bd3baf6: Status 404 returned error can't find the container with id 4c08efe1a2fb3f7a54cfcfbb81890fb6139c1569d2dc8826629dc5d40bd3baf6 Oct 01 16:15:55 crc kubenswrapper[4726]: I1001 16:15:55.955720 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-1b50-account-create-qw2dx" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.159632 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6bf9-account-create-9xm7t"] Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.357446 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6bf9-account-create-9xm7t" event={"ID":"1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4","Type":"ContainerStarted","Data":"d762290dd72ba2a184a8f94ae306eb60c39157faa1308aad41bdcf2b8ea60508"} Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.359146 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6f33-account-create-rw9w7" event={"ID":"50c724bf-9bf3-4e8b-97c5-09199a794376","Type":"ContainerDied","Data":"fe7ba6a0250898750a90261e74412b10d5968470b6ca2000c9aac04fcfeba241"} Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.359173 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe7ba6a0250898750a90261e74412b10d5968470b6ca2000c9aac04fcfeba241" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.359220 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6f33-account-create-rw9w7" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.362645 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7b0c-account-create-jf4fq" event={"ID":"56bde538-7fa3-4a9b-be9f-87ee018e6a43","Type":"ContainerStarted","Data":"4c08efe1a2fb3f7a54cfcfbb81890fb6139c1569d2dc8826629dc5d40bd3baf6"} Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.398368 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-1b50-account-create-qw2dx"] Oct 01 16:15:56 crc kubenswrapper[4726]: W1001 16:15:56.404347 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod708ea401_1d87_4a51_ab33_b9318fe542e2.slice/crio-57625e41812a22a1d6eb572c6396af6f3bb79e04ef4633aa2a1626e3e857636f WatchSource:0}: Error finding container 57625e41812a22a1d6eb572c6396af6f3bb79e04ef4633aa2a1626e3e857636f: Status 404 returned error can't find the container with id 57625e41812a22a1d6eb572c6396af6f3bb79e04ef4633aa2a1626e3e857636f Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.711747 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.750552 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8kpg\" (UniqueName: \"kubernetes.io/projected/602d1e20-fa25-4183-9801-abadbbbe3495-kube-api-access-q8kpg\") pod \"602d1e20-fa25-4183-9801-abadbbbe3495\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.750629 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-log-ovn\") pod \"602d1e20-fa25-4183-9801-abadbbbe3495\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.750659 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/602d1e20-fa25-4183-9801-abadbbbe3495-additional-scripts\") pod \"602d1e20-fa25-4183-9801-abadbbbe3495\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.750729 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/602d1e20-fa25-4183-9801-abadbbbe3495-scripts\") pod \"602d1e20-fa25-4183-9801-abadbbbe3495\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.750749 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "602d1e20-fa25-4183-9801-abadbbbe3495" (UID: "602d1e20-fa25-4183-9801-abadbbbe3495"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.750825 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-run-ovn\") pod \"602d1e20-fa25-4183-9801-abadbbbe3495\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.750863 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-run\") pod \"602d1e20-fa25-4183-9801-abadbbbe3495\" (UID: \"602d1e20-fa25-4183-9801-abadbbbe3495\") " Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.751183 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "602d1e20-fa25-4183-9801-abadbbbe3495" (UID: "602d1e20-fa25-4183-9801-abadbbbe3495"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.751297 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-run" (OuterVolumeSpecName: "var-run") pod "602d1e20-fa25-4183-9801-abadbbbe3495" (UID: "602d1e20-fa25-4183-9801-abadbbbe3495"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.751476 4726 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.751551 4726 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.751491 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/602d1e20-fa25-4183-9801-abadbbbe3495-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "602d1e20-fa25-4183-9801-abadbbbe3495" (UID: "602d1e20-fa25-4183-9801-abadbbbe3495"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.751734 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/602d1e20-fa25-4183-9801-abadbbbe3495-scripts" (OuterVolumeSpecName: "scripts") pod "602d1e20-fa25-4183-9801-abadbbbe3495" (UID: "602d1e20-fa25-4183-9801-abadbbbe3495"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.759273 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/602d1e20-fa25-4183-9801-abadbbbe3495-kube-api-access-q8kpg" (OuterVolumeSpecName: "kube-api-access-q8kpg") pod "602d1e20-fa25-4183-9801-abadbbbe3495" (UID: "602d1e20-fa25-4183-9801-abadbbbe3495"). InnerVolumeSpecName "kube-api-access-q8kpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.852560 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8kpg\" (UniqueName: \"kubernetes.io/projected/602d1e20-fa25-4183-9801-abadbbbe3495-kube-api-access-q8kpg\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.852603 4726 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/602d1e20-fa25-4183-9801-abadbbbe3495-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.852613 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/602d1e20-fa25-4183-9801-abadbbbe3495-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:56 crc kubenswrapper[4726]: I1001 16:15:56.852621 4726 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/602d1e20-fa25-4183-9801-abadbbbe3495-var-run\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.370122 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-1b50-account-create-qw2dx" event={"ID":"708ea401-1d87-4a51-ab33-b9318fe542e2","Type":"ContainerStarted","Data":"57625e41812a22a1d6eb572c6396af6f3bb79e04ef4633aa2a1626e3e857636f"} Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.372323 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7b0c-account-create-jf4fq" event={"ID":"56bde538-7fa3-4a9b-be9f-87ee018e6a43","Type":"ContainerStarted","Data":"14333894a4ee2c149a52cce8efff90eba2703e7d963369d3b26cc6f56a1f2c0d"} Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.374039 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z4lk4-config-qmmd8" event={"ID":"602d1e20-fa25-4183-9801-abadbbbe3495","Type":"ContainerDied","Data":"1a88cbef062fb21a5ef20e55ef413499e2b023c1f21493626823013d56c23cbc"} Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.374078 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a88cbef062fb21a5ef20e55ef413499e2b023c1f21493626823013d56c23cbc" Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.374133 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z4lk4-config-qmmd8" Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.848161 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-z4lk4-config-qmmd8"] Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.867250 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-z4lk4-config-qmmd8"] Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.897652 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-z4lk4" Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.919797 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-zwfrl"] Oct 01 16:15:57 crc kubenswrapper[4726]: E1001 16:15:57.920178 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="602d1e20-fa25-4183-9801-abadbbbe3495" containerName="ovn-config" Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.920195 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="602d1e20-fa25-4183-9801-abadbbbe3495" containerName="ovn-config" Oct 01 16:15:57 crc kubenswrapper[4726]: E1001 16:15:57.920212 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c724bf-9bf3-4e8b-97c5-09199a794376" containerName="mariadb-account-create" Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.920218 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c724bf-9bf3-4e8b-97c5-09199a794376" containerName="mariadb-account-create" Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.920382 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="602d1e20-fa25-4183-9801-abadbbbe3495" containerName="ovn-config" Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.920398 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="50c724bf-9bf3-4e8b-97c5-09199a794376" containerName="mariadb-account-create" Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.920904 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zwfrl" Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.923084 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.923295 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-7lwk2" Oct 01 16:15:57 crc kubenswrapper[4726]: I1001 16:15:57.935574 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-zwfrl"] Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.074720 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-db-sync-config-data\") pod \"glance-db-sync-zwfrl\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " pod="openstack/glance-db-sync-zwfrl" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.074801 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-config-data\") pod \"glance-db-sync-zwfrl\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " pod="openstack/glance-db-sync-zwfrl" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.074851 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-combined-ca-bundle\") pod \"glance-db-sync-zwfrl\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " pod="openstack/glance-db-sync-zwfrl" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.074879 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7pjv\" (UniqueName: \"kubernetes.io/projected/bdf2f47d-bf60-4b49-ae78-1b7823838797-kube-api-access-f7pjv\") pod \"glance-db-sync-zwfrl\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " pod="openstack/glance-db-sync-zwfrl" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.176188 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-db-sync-config-data\") pod \"glance-db-sync-zwfrl\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " pod="openstack/glance-db-sync-zwfrl" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.176287 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-config-data\") pod \"glance-db-sync-zwfrl\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " pod="openstack/glance-db-sync-zwfrl" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.176350 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-combined-ca-bundle\") pod \"glance-db-sync-zwfrl\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " pod="openstack/glance-db-sync-zwfrl" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.176382 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7pjv\" (UniqueName: \"kubernetes.io/projected/bdf2f47d-bf60-4b49-ae78-1b7823838797-kube-api-access-f7pjv\") pod \"glance-db-sync-zwfrl\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " pod="openstack/glance-db-sync-zwfrl" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.182656 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-combined-ca-bundle\") pod \"glance-db-sync-zwfrl\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " pod="openstack/glance-db-sync-zwfrl" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.182724 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-config-data\") pod \"glance-db-sync-zwfrl\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " pod="openstack/glance-db-sync-zwfrl" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.182741 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-db-sync-config-data\") pod \"glance-db-sync-zwfrl\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " pod="openstack/glance-db-sync-zwfrl" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.191580 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7pjv\" (UniqueName: \"kubernetes.io/projected/bdf2f47d-bf60-4b49-ae78-1b7823838797-kube-api-access-f7pjv\") pod \"glance-db-sync-zwfrl\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " pod="openstack/glance-db-sync-zwfrl" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.275213 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zwfrl" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.393885 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6bf9-account-create-9xm7t" event={"ID":"1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4","Type":"ContainerStarted","Data":"48cb20845a5026054f33727a088ddf596c76343f7fbe91163c232e0c334762dc"} Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.396130 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-1b50-account-create-qw2dx" event={"ID":"708ea401-1d87-4a51-ab33-b9318fe542e2","Type":"ContainerStarted","Data":"ecb0b0dac1d3a368c9250b62f6fee94ee5de5548750f72d5da165556dd8b3c5e"} Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.420210 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-6bf9-account-create-9xm7t" podStartSLOduration=3.420187235 podStartE2EDuration="3.420187235s" podCreationTimestamp="2025-10-01 16:15:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:15:58.411443631 +0000 UTC m=+1011.312996208" watchObservedRunningTime="2025-10-01 16:15:58.420187235 +0000 UTC m=+1011.321739812" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.435747 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-7b0c-account-create-jf4fq" podStartSLOduration=3.435732187 podStartE2EDuration="3.435732187s" podCreationTimestamp="2025-10-01 16:15:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:15:58.424631924 +0000 UTC m=+1011.326184521" watchObservedRunningTime="2025-10-01 16:15:58.435732187 +0000 UTC m=+1011.337284754" Oct 01 16:15:58 crc kubenswrapper[4726]: I1001 16:15:58.814785 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-zwfrl"] Oct 01 16:15:59 crc kubenswrapper[4726]: I1001 16:15:59.412537 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zwfrl" event={"ID":"bdf2f47d-bf60-4b49-ae78-1b7823838797","Type":"ContainerStarted","Data":"69d68905f8bc2dfd35245d7fb80d3858575b8d612170ff741c6a27183cb3f9e2"} Oct 01 16:15:59 crc kubenswrapper[4726]: I1001 16:15:59.437827 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-1b50-account-create-qw2dx" podStartSLOduration=4.437796157 podStartE2EDuration="4.437796157s" podCreationTimestamp="2025-10-01 16:15:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:15:59.434388228 +0000 UTC m=+1012.335940805" watchObservedRunningTime="2025-10-01 16:15:59.437796157 +0000 UTC m=+1012.339348774" Oct 01 16:15:59 crc kubenswrapper[4726]: I1001 16:15:59.826873 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="602d1e20-fa25-4183-9801-abadbbbe3495" path="/var/lib/kubelet/pods/602d1e20-fa25-4183-9801-abadbbbe3495/volumes" Oct 01 16:16:04 crc kubenswrapper[4726]: I1001 16:16:04.516973 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:16:04 crc kubenswrapper[4726]: I1001 16:16:04.529540 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d192df1-7269-4493-b17b-4037addc226e-etc-swift\") pod \"swift-storage-0\" (UID: \"6d192df1-7269-4493-b17b-4037addc226e\") " pod="openstack/swift-storage-0" Oct 01 16:16:04 crc kubenswrapper[4726]: I1001 16:16:04.665608 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 01 16:16:07 crc kubenswrapper[4726]: I1001 16:16:07.165428 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 01 16:16:07 crc kubenswrapper[4726]: W1001 16:16:07.246002 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d192df1_7269_4493_b17b_4037addc226e.slice/crio-c390d7e6a410d64600222948a73f879b9ddc219360649c0022f1aef3bf8c15d9 WatchSource:0}: Error finding container c390d7e6a410d64600222948a73f879b9ddc219360649c0022f1aef3bf8c15d9: Status 404 returned error can't find the container with id c390d7e6a410d64600222948a73f879b9ddc219360649c0022f1aef3bf8c15d9 Oct 01 16:16:07 crc kubenswrapper[4726]: I1001 16:16:07.506444 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"c390d7e6a410d64600222948a73f879b9ddc219360649c0022f1aef3bf8c15d9"} Oct 01 16:16:07 crc kubenswrapper[4726]: I1001 16:16:07.509379 4726 generic.go:334] "Generic (PLEG): container finished" podID="56bde538-7fa3-4a9b-be9f-87ee018e6a43" containerID="14333894a4ee2c149a52cce8efff90eba2703e7d963369d3b26cc6f56a1f2c0d" exitCode=0 Oct 01 16:16:07 crc kubenswrapper[4726]: I1001 16:16:07.509409 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7b0c-account-create-jf4fq" event={"ID":"56bde538-7fa3-4a9b-be9f-87ee018e6a43","Type":"ContainerDied","Data":"14333894a4ee2c149a52cce8efff90eba2703e7d963369d3b26cc6f56a1f2c0d"} Oct 01 16:16:08 crc kubenswrapper[4726]: E1001 16:16:08.000964 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-keystone:current-podified" Oct 01 16:16:08 crc kubenswrapper[4726]: E1001 16:16:08.001521 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:keystone-db-sync,Image:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,Command:[/bin/bash],Args:[-c keystone-manage db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/keystone/keystone.conf,SubPath:keystone.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-87hgb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42425,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42425,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-db-sync-2df5f_openstack(4a875fd8-855b-48b4-91ad-043ad70711a1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:16:08 crc kubenswrapper[4726]: E1001 16:16:08.002844 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/keystone-db-sync-2df5f" podUID="4a875fd8-855b-48b4-91ad-043ad70711a1" Oct 01 16:16:08 crc kubenswrapper[4726]: I1001 16:16:08.519145 4726 generic.go:334] "Generic (PLEG): container finished" podID="1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4" containerID="48cb20845a5026054f33727a088ddf596c76343f7fbe91163c232e0c334762dc" exitCode=0 Oct 01 16:16:08 crc kubenswrapper[4726]: I1001 16:16:08.519239 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6bf9-account-create-9xm7t" event={"ID":"1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4","Type":"ContainerDied","Data":"48cb20845a5026054f33727a088ddf596c76343f7fbe91163c232e0c334762dc"} Oct 01 16:16:08 crc kubenswrapper[4726]: E1001 16:16:08.520604 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-keystone:current-podified\\\"\"" pod="openstack/keystone-db-sync-2df5f" podUID="4a875fd8-855b-48b4-91ad-043ad70711a1" Oct 01 16:16:08 crc kubenswrapper[4726]: I1001 16:16:08.968018 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7b0c-account-create-jf4fq" Oct 01 16:16:09 crc kubenswrapper[4726]: I1001 16:16:09.103123 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qdbv\" (UniqueName: \"kubernetes.io/projected/56bde538-7fa3-4a9b-be9f-87ee018e6a43-kube-api-access-5qdbv\") pod \"56bde538-7fa3-4a9b-be9f-87ee018e6a43\" (UID: \"56bde538-7fa3-4a9b-be9f-87ee018e6a43\") " Oct 01 16:16:09 crc kubenswrapper[4726]: I1001 16:16:09.111930 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56bde538-7fa3-4a9b-be9f-87ee018e6a43-kube-api-access-5qdbv" (OuterVolumeSpecName: "kube-api-access-5qdbv") pod "56bde538-7fa3-4a9b-be9f-87ee018e6a43" (UID: "56bde538-7fa3-4a9b-be9f-87ee018e6a43"). InnerVolumeSpecName "kube-api-access-5qdbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:16:09 crc kubenswrapper[4726]: I1001 16:16:09.204938 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qdbv\" (UniqueName: \"kubernetes.io/projected/56bde538-7fa3-4a9b-be9f-87ee018e6a43-kube-api-access-5qdbv\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:09 crc kubenswrapper[4726]: I1001 16:16:09.533746 4726 generic.go:334] "Generic (PLEG): container finished" podID="708ea401-1d87-4a51-ab33-b9318fe542e2" containerID="ecb0b0dac1d3a368c9250b62f6fee94ee5de5548750f72d5da165556dd8b3c5e" exitCode=0 Oct 01 16:16:09 crc kubenswrapper[4726]: I1001 16:16:09.533838 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-1b50-account-create-qw2dx" event={"ID":"708ea401-1d87-4a51-ab33-b9318fe542e2","Type":"ContainerDied","Data":"ecb0b0dac1d3a368c9250b62f6fee94ee5de5548750f72d5da165556dd8b3c5e"} Oct 01 16:16:09 crc kubenswrapper[4726]: I1001 16:16:09.536027 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7b0c-account-create-jf4fq" event={"ID":"56bde538-7fa3-4a9b-be9f-87ee018e6a43","Type":"ContainerDied","Data":"4c08efe1a2fb3f7a54cfcfbb81890fb6139c1569d2dc8826629dc5d40bd3baf6"} Oct 01 16:16:09 crc kubenswrapper[4726]: I1001 16:16:09.536113 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c08efe1a2fb3f7a54cfcfbb81890fb6139c1569d2dc8826629dc5d40bd3baf6" Oct 01 16:16:09 crc kubenswrapper[4726]: I1001 16:16:09.536039 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7b0c-account-create-jf4fq" Oct 01 16:16:09 crc kubenswrapper[4726]: I1001 16:16:09.894617 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6bf9-account-create-9xm7t" Oct 01 16:16:09 crc kubenswrapper[4726]: I1001 16:16:09.924006 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99424\" (UniqueName: \"kubernetes.io/projected/1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4-kube-api-access-99424\") pod \"1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4\" (UID: \"1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4\") " Oct 01 16:16:09 crc kubenswrapper[4726]: I1001 16:16:09.930491 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4-kube-api-access-99424" (OuterVolumeSpecName: "kube-api-access-99424") pod "1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4" (UID: "1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4"). InnerVolumeSpecName "kube-api-access-99424". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:16:10 crc kubenswrapper[4726]: I1001 16:16:10.026105 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99424\" (UniqueName: \"kubernetes.io/projected/1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4-kube-api-access-99424\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:10 crc kubenswrapper[4726]: I1001 16:16:10.550681 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6bf9-account-create-9xm7t" event={"ID":"1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4","Type":"ContainerDied","Data":"d762290dd72ba2a184a8f94ae306eb60c39157faa1308aad41bdcf2b8ea60508"} Oct 01 16:16:10 crc kubenswrapper[4726]: I1001 16:16:10.550746 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d762290dd72ba2a184a8f94ae306eb60c39157faa1308aad41bdcf2b8ea60508" Oct 01 16:16:10 crc kubenswrapper[4726]: I1001 16:16:10.550764 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6bf9-account-create-9xm7t" Oct 01 16:16:19 crc kubenswrapper[4726]: I1001 16:16:19.319270 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-1b50-account-create-qw2dx" Oct 01 16:16:19 crc kubenswrapper[4726]: I1001 16:16:19.358362 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7jh8\" (UniqueName: \"kubernetes.io/projected/708ea401-1d87-4a51-ab33-b9318fe542e2-kube-api-access-w7jh8\") pod \"708ea401-1d87-4a51-ab33-b9318fe542e2\" (UID: \"708ea401-1d87-4a51-ab33-b9318fe542e2\") " Oct 01 16:16:19 crc kubenswrapper[4726]: I1001 16:16:19.363692 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/708ea401-1d87-4a51-ab33-b9318fe542e2-kube-api-access-w7jh8" (OuterVolumeSpecName: "kube-api-access-w7jh8") pod "708ea401-1d87-4a51-ab33-b9318fe542e2" (UID: "708ea401-1d87-4a51-ab33-b9318fe542e2"). InnerVolumeSpecName "kube-api-access-w7jh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:16:19 crc kubenswrapper[4726]: I1001 16:16:19.460235 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7jh8\" (UniqueName: \"kubernetes.io/projected/708ea401-1d87-4a51-ab33-b9318fe542e2-kube-api-access-w7jh8\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:19 crc kubenswrapper[4726]: I1001 16:16:19.629151 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-1b50-account-create-qw2dx" event={"ID":"708ea401-1d87-4a51-ab33-b9318fe542e2","Type":"ContainerDied","Data":"57625e41812a22a1d6eb572c6396af6f3bb79e04ef4633aa2a1626e3e857636f"} Oct 01 16:16:19 crc kubenswrapper[4726]: I1001 16:16:19.629185 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-1b50-account-create-qw2dx" Oct 01 16:16:19 crc kubenswrapper[4726]: I1001 16:16:19.629202 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57625e41812a22a1d6eb572c6396af6f3bb79e04ef4633aa2a1626e3e857636f" Oct 01 16:16:26 crc kubenswrapper[4726]: E1001 16:16:26.110189 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Oct 01 16:16:26 crc kubenswrapper[4726]: E1001 16:16:26.110986 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f7pjv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-zwfrl_openstack(bdf2f47d-bf60-4b49-ae78-1b7823838797): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:16:26 crc kubenswrapper[4726]: E1001 16:16:26.112245 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-zwfrl" podUID="bdf2f47d-bf60-4b49-ae78-1b7823838797" Oct 01 16:16:26 crc kubenswrapper[4726]: I1001 16:16:26.688229 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"03c700fcd7954af66327750048e8d2ba038da66b6d5dfad083d5dfbf275a3956"} Oct 01 16:16:26 crc kubenswrapper[4726]: I1001 16:16:26.691153 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-2df5f" event={"ID":"4a875fd8-855b-48b4-91ad-043ad70711a1","Type":"ContainerStarted","Data":"49212a1bbb2d4fb57572e6e200ff3a2d957958def82921f59307fad9a1fde616"} Oct 01 16:16:26 crc kubenswrapper[4726]: E1001 16:16:26.691889 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-zwfrl" podUID="bdf2f47d-bf60-4b49-ae78-1b7823838797" Oct 01 16:16:26 crc kubenswrapper[4726]: I1001 16:16:26.727313 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-2df5f" podStartSLOduration=2.238655936 podStartE2EDuration="34.727289225s" podCreationTimestamp="2025-10-01 16:15:52 +0000 UTC" firstStartedPulling="2025-10-01 16:15:53.679148242 +0000 UTC m=+1006.580700819" lastFinishedPulling="2025-10-01 16:16:26.167781491 +0000 UTC m=+1039.069334108" observedRunningTime="2025-10-01 16:16:26.725267567 +0000 UTC m=+1039.626820154" watchObservedRunningTime="2025-10-01 16:16:26.727289225 +0000 UTC m=+1039.628841802" Oct 01 16:16:27 crc kubenswrapper[4726]: I1001 16:16:27.704200 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"9409d9ebd08d6e7904a4f8ddce4fe7522961f6252bf7310c47ddb92565a05f62"} Oct 01 16:16:27 crc kubenswrapper[4726]: I1001 16:16:27.704250 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"7f9126b382f08ba473e8d9dda712f35651844268bbcefba0b4b861bab8547b8a"} Oct 01 16:16:27 crc kubenswrapper[4726]: I1001 16:16:27.704262 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"fbb662642238ec7c4464c4ba49cec4abcd7047198ba2562243a83149b2b6a0d5"} Oct 01 16:16:28 crc kubenswrapper[4726]: I1001 16:16:28.721114 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"dcd659c0676cef0afff237bbd738ee5ab998292672173163203b1201d4408b7a"} Oct 01 16:16:29 crc kubenswrapper[4726]: I1001 16:16:29.731388 4726 generic.go:334] "Generic (PLEG): container finished" podID="4a875fd8-855b-48b4-91ad-043ad70711a1" containerID="49212a1bbb2d4fb57572e6e200ff3a2d957958def82921f59307fad9a1fde616" exitCode=0 Oct 01 16:16:29 crc kubenswrapper[4726]: I1001 16:16:29.731498 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-2df5f" event={"ID":"4a875fd8-855b-48b4-91ad-043ad70711a1","Type":"ContainerDied","Data":"49212a1bbb2d4fb57572e6e200ff3a2d957958def82921f59307fad9a1fde616"} Oct 01 16:16:29 crc kubenswrapper[4726]: I1001 16:16:29.736374 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"11953a99825f99735563b7fc2f6ac3e427263f5c5a44533dec1762136e91b863"} Oct 01 16:16:29 crc kubenswrapper[4726]: I1001 16:16:29.736403 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"d48ae510698324f5a743033538a5bc44dafbb5538fe822291fed76b713d82f1d"} Oct 01 16:16:29 crc kubenswrapper[4726]: I1001 16:16:29.736414 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"f3723931738365b31719d9cbcd1228ed21e0589d9484402535f4485787fe9f0e"} Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.180499 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-2df5f" Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.363799 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a875fd8-855b-48b4-91ad-043ad70711a1-config-data\") pod \"4a875fd8-855b-48b4-91ad-043ad70711a1\" (UID: \"4a875fd8-855b-48b4-91ad-043ad70711a1\") " Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.363861 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a875fd8-855b-48b4-91ad-043ad70711a1-combined-ca-bundle\") pod \"4a875fd8-855b-48b4-91ad-043ad70711a1\" (UID: \"4a875fd8-855b-48b4-91ad-043ad70711a1\") " Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.364039 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87hgb\" (UniqueName: \"kubernetes.io/projected/4a875fd8-855b-48b4-91ad-043ad70711a1-kube-api-access-87hgb\") pod \"4a875fd8-855b-48b4-91ad-043ad70711a1\" (UID: \"4a875fd8-855b-48b4-91ad-043ad70711a1\") " Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.371461 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a875fd8-855b-48b4-91ad-043ad70711a1-kube-api-access-87hgb" (OuterVolumeSpecName: "kube-api-access-87hgb") pod "4a875fd8-855b-48b4-91ad-043ad70711a1" (UID: "4a875fd8-855b-48b4-91ad-043ad70711a1"). InnerVolumeSpecName "kube-api-access-87hgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.399319 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a875fd8-855b-48b4-91ad-043ad70711a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a875fd8-855b-48b4-91ad-043ad70711a1" (UID: "4a875fd8-855b-48b4-91ad-043ad70711a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.425961 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a875fd8-855b-48b4-91ad-043ad70711a1-config-data" (OuterVolumeSpecName: "config-data") pod "4a875fd8-855b-48b4-91ad-043ad70711a1" (UID: "4a875fd8-855b-48b4-91ad-043ad70711a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.465332 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87hgb\" (UniqueName: \"kubernetes.io/projected/4a875fd8-855b-48b4-91ad-043ad70711a1-kube-api-access-87hgb\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.465372 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a875fd8-855b-48b4-91ad-043ad70711a1-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.465389 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a875fd8-855b-48b4-91ad-043ad70711a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.765237 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"0e813ed6438065a372ad7dd0136b2e7a86ae6a70a2c3096f54bf9bf83caea15d"} Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.765856 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"4bd38b996cbb9127bbf2d0d744303ad11d41d81875a29ba1ecf0cff912bf8501"} Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.765871 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"1549c57590e3a151caec74c157dacb96552372f5077aea11aeaa60225bf7e11e"} Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.765899 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"0abd1034bb9c8d336f50e40e3c076633e426dbd82af894de9655fb961e5ddc37"} Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.765911 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"5c3983c85095d76b1050527a2c013a3f228e55b9cbe4367d75bc104387bd3e43"} Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.766946 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-2df5f" event={"ID":"4a875fd8-855b-48b4-91ad-043ad70711a1","Type":"ContainerDied","Data":"33a6c1e135279a195ed32f2aab1ff9a5ae4c860960fd388fd910b5cba8cb8e52"} Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.766970 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33a6c1e135279a195ed32f2aab1ff9a5ae4c860960fd388fd910b5cba8cb8e52" Oct 01 16:16:31 crc kubenswrapper[4726]: I1001 16:16:31.767114 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-2df5f" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.046538 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-m7bvd"] Oct 01 16:16:32 crc kubenswrapper[4726]: E1001 16:16:32.046922 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="708ea401-1d87-4a51-ab33-b9318fe542e2" containerName="mariadb-account-create" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.046941 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="708ea401-1d87-4a51-ab33-b9318fe542e2" containerName="mariadb-account-create" Oct 01 16:16:32 crc kubenswrapper[4726]: E1001 16:16:32.046975 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4" containerName="mariadb-account-create" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.046984 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4" containerName="mariadb-account-create" Oct 01 16:16:32 crc kubenswrapper[4726]: E1001 16:16:32.047007 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56bde538-7fa3-4a9b-be9f-87ee018e6a43" containerName="mariadb-account-create" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.047015 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="56bde538-7fa3-4a9b-be9f-87ee018e6a43" containerName="mariadb-account-create" Oct 01 16:16:32 crc kubenswrapper[4726]: E1001 16:16:32.047025 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a875fd8-855b-48b4-91ad-043ad70711a1" containerName="keystone-db-sync" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.047033 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a875fd8-855b-48b4-91ad-043ad70711a1" containerName="keystone-db-sync" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.047239 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a875fd8-855b-48b4-91ad-043ad70711a1" containerName="keystone-db-sync" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.047268 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="56bde538-7fa3-4a9b-be9f-87ee018e6a43" containerName="mariadb-account-create" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.047291 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4" containerName="mariadb-account-create" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.047316 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="708ea401-1d87-4a51-ab33-b9318fe542e2" containerName="mariadb-account-create" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.048274 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.064134 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-zwwsm"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.065014 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.068239 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.068450 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.068689 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5q5v6" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.068810 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.073666 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-m7bvd"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.135933 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zwwsm"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.181934 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-config\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.182835 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.182929 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-scripts\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.183906 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.184018 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-config-data\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.184099 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-fernet-keys\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.184164 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-combined-ca-bundle\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.184289 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2tqk\" (UniqueName: \"kubernetes.io/projected/a8c3496b-70c5-4e68-8781-b4936a302273-kube-api-access-d2tqk\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.184360 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w58w\" (UniqueName: \"kubernetes.io/projected/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-kube-api-access-7w58w\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.184443 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.185085 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-credential-keys\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.195226 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7945d7f85f-2qfqm"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.199604 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.204097 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.204948 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.205342 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-rvpbr" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.205496 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.234888 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7945d7f85f-2qfqm"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.287480 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.287525 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-scripts\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.287555 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.287595 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-config-data\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.287609 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-fernet-keys\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.287628 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-combined-ca-bundle\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.287670 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2tqk\" (UniqueName: \"kubernetes.io/projected/a8c3496b-70c5-4e68-8781-b4936a302273-kube-api-access-d2tqk\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.287688 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w58w\" (UniqueName: \"kubernetes.io/projected/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-kube-api-access-7w58w\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.287712 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.287733 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-credential-keys\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.287775 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-config\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.288524 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-config\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.289006 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.292803 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.293774 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.297464 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-combined-ca-bundle\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.297612 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-scripts\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.305561 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-credential-keys\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.313605 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-fc7f8d489-j7grp"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.317103 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.318859 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-fernet-keys\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.334161 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w58w\" (UniqueName: \"kubernetes.io/projected/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-kube-api-access-7w58w\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.336676 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2tqk\" (UniqueName: \"kubernetes.io/projected/a8c3496b-70c5-4e68-8781-b4936a302273-kube-api-access-d2tqk\") pod \"dnsmasq-dns-5c9d85d47c-m7bvd\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.343151 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-config-data\") pod \"keystone-bootstrap-zwwsm\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.364992 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-fc7f8d489-j7grp"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.379798 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.388300 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.390348 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a2911957-529c-46a4-8b63-e98beb36809b-horizon-secret-key\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.390374 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w22z\" (UniqueName: \"kubernetes.io/projected/62a37356-6f11-4c75-a452-570a847a0518-kube-api-access-2w22z\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.390416 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2911957-529c-46a4-8b63-e98beb36809b-config-data\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.390443 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/62a37356-6f11-4c75-a452-570a847a0518-horizon-secret-key\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.390485 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a2911957-529c-46a4-8b63-e98beb36809b-scripts\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.390501 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62a37356-6f11-4c75-a452-570a847a0518-config-data\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.390517 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62a37356-6f11-4c75-a452-570a847a0518-scripts\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.390535 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62a37356-6f11-4c75-a452-570a847a0518-logs\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.390571 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2911957-529c-46a4-8b63-e98beb36809b-logs\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.390596 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w5s9\" (UniqueName: \"kubernetes.io/projected/a2911957-529c-46a4-8b63-e98beb36809b-kube-api-access-7w5s9\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.431190 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-skmf7"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.432262 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-skmf7" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.436264 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.436597 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-pdhpb" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.436779 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.478354 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-gsljc"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.479434 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.490705 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-x9fkn" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.491091 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.491279 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.502001 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-skmf7"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505316 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q72v4\" (UniqueName: \"kubernetes.io/projected/7b994ac8-f28c-458a-873f-4270e9f0f36b-kube-api-access-q72v4\") pod \"neutron-db-sync-skmf7\" (UID: \"7b994ac8-f28c-458a-873f-4270e9f0f36b\") " pod="openstack/neutron-db-sync-skmf7" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505422 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znxm2\" (UniqueName: \"kubernetes.io/projected/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-kube-api-access-znxm2\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505457 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2911957-529c-46a4-8b63-e98beb36809b-logs\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505476 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7b994ac8-f28c-458a-873f-4270e9f0f36b-config\") pod \"neutron-db-sync-skmf7\" (UID: \"7b994ac8-f28c-458a-873f-4270e9f0f36b\") " pod="openstack/neutron-db-sync-skmf7" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505509 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-db-sync-config-data\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505536 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w5s9\" (UniqueName: \"kubernetes.io/projected/a2911957-529c-46a4-8b63-e98beb36809b-kube-api-access-7w5s9\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505561 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a2911957-529c-46a4-8b63-e98beb36809b-horizon-secret-key\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505580 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w22z\" (UniqueName: \"kubernetes.io/projected/62a37356-6f11-4c75-a452-570a847a0518-kube-api-access-2w22z\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505598 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-scripts\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505637 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-etc-machine-id\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505671 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-combined-ca-bundle\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505696 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2911957-529c-46a4-8b63-e98beb36809b-config-data\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505721 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b994ac8-f28c-458a-873f-4270e9f0f36b-combined-ca-bundle\") pod \"neutron-db-sync-skmf7\" (UID: \"7b994ac8-f28c-458a-873f-4270e9f0f36b\") " pod="openstack/neutron-db-sync-skmf7" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505750 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/62a37356-6f11-4c75-a452-570a847a0518-horizon-secret-key\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505807 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-config-data\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505842 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a2911957-529c-46a4-8b63-e98beb36809b-scripts\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505866 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62a37356-6f11-4c75-a452-570a847a0518-config-data\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505881 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62a37356-6f11-4c75-a452-570a847a0518-scripts\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.505905 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62a37356-6f11-4c75-a452-570a847a0518-logs\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.506684 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2911957-529c-46a4-8b63-e98beb36809b-logs\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.507613 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a2911957-529c-46a4-8b63-e98beb36809b-scripts\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.508066 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62a37356-6f11-4c75-a452-570a847a0518-logs\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.510875 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2911957-529c-46a4-8b63-e98beb36809b-config-data\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.512599 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a2911957-529c-46a4-8b63-e98beb36809b-horizon-secret-key\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.513140 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62a37356-6f11-4c75-a452-570a847a0518-config-data\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.515537 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62a37356-6f11-4c75-a452-570a847a0518-scripts\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.526076 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/62a37356-6f11-4c75-a452-570a847a0518-horizon-secret-key\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.537922 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w5s9\" (UniqueName: \"kubernetes.io/projected/a2911957-529c-46a4-8b63-e98beb36809b-kube-api-access-7w5s9\") pod \"horizon-7945d7f85f-2qfqm\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.537995 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-gsljc"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.543704 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.561648 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w22z\" (UniqueName: \"kubernetes.io/projected/62a37356-6f11-4c75-a452-570a847a0518-kube-api-access-2w22z\") pod \"horizon-fc7f8d489-j7grp\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.580041 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-bhxd2"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.581290 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bhxd2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.586416 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.586478 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-f2dgd" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.607783 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q72v4\" (UniqueName: \"kubernetes.io/projected/7b994ac8-f28c-458a-873f-4270e9f0f36b-kube-api-access-q72v4\") pod \"neutron-db-sync-skmf7\" (UID: \"7b994ac8-f28c-458a-873f-4270e9f0f36b\") " pod="openstack/neutron-db-sync-skmf7" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.607841 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znxm2\" (UniqueName: \"kubernetes.io/projected/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-kube-api-access-znxm2\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.607867 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7b994ac8-f28c-458a-873f-4270e9f0f36b-config\") pod \"neutron-db-sync-skmf7\" (UID: \"7b994ac8-f28c-458a-873f-4270e9f0f36b\") " pod="openstack/neutron-db-sync-skmf7" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.607889 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-db-sync-config-data\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.607913 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-scripts\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.607942 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-etc-machine-id\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.607964 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-combined-ca-bundle\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.607989 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b994ac8-f28c-458a-873f-4270e9f0f36b-combined-ca-bundle\") pod \"neutron-db-sync-skmf7\" (UID: \"7b994ac8-f28c-458a-873f-4270e9f0f36b\") " pod="openstack/neutron-db-sync-skmf7" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.608027 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-config-data\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.611062 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-etc-machine-id\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.617691 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-db-sync-config-data\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.624880 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-bhxd2"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.630863 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-config-data\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.638030 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.640297 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.641930 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b994ac8-f28c-458a-873f-4270e9f0f36b-combined-ca-bundle\") pod \"neutron-db-sync-skmf7\" (UID: \"7b994ac8-f28c-458a-873f-4270e9f0f36b\") " pod="openstack/neutron-db-sync-skmf7" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.642610 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-scripts\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.643939 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.646728 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.654949 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q72v4\" (UniqueName: \"kubernetes.io/projected/7b994ac8-f28c-458a-873f-4270e9f0f36b-kube-api-access-q72v4\") pod \"neutron-db-sync-skmf7\" (UID: \"7b994ac8-f28c-458a-873f-4270e9f0f36b\") " pod="openstack/neutron-db-sync-skmf7" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.655690 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7b994ac8-f28c-458a-873f-4270e9f0f36b-config\") pod \"neutron-db-sync-skmf7\" (UID: \"7b994ac8-f28c-458a-873f-4270e9f0f36b\") " pod="openstack/neutron-db-sync-skmf7" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.656198 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-combined-ca-bundle\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.665703 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znxm2\" (UniqueName: \"kubernetes.io/projected/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-kube-api-access-znxm2\") pod \"cinder-db-sync-gsljc\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.672576 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.681914 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-ks2wk"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.685507 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.690668 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.691573 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-fvv6p" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.691613 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.705582 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ks2wk"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.709926 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-combined-ca-bundle\") pod \"barbican-db-sync-bhxd2\" (UID: \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\") " pod="openstack/barbican-db-sync-bhxd2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.709976 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd4s2\" (UniqueName: \"kubernetes.io/projected/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-kube-api-access-hd4s2\") pod \"barbican-db-sync-bhxd2\" (UID: \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\") " pod="openstack/barbican-db-sync-bhxd2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.710037 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-db-sync-config-data\") pod \"barbican-db-sync-bhxd2\" (UID: \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\") " pod="openstack/barbican-db-sync-bhxd2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.725828 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-m7bvd"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.734737 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-jvrp2"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.736458 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.740760 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.746391 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-jvrp2"] Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.779417 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-skmf7" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.801456 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"d634d077c7a6228a3e2f4c2b13c8ed2bf3b3b12e632d198c1557f7df438b34e8"} Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.801503 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6d192df1-7269-4493-b17b-4037addc226e","Type":"ContainerStarted","Data":"a9c76fe47e1284759b1c348d6f35ac7cddc09309efb463b90f806907c85c7993"} Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813007 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjd46\" (UniqueName: \"kubernetes.io/projected/395cb84c-684a-48ec-aa24-03f5a7f218ff-kube-api-access-kjd46\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813046 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-combined-ca-bundle\") pod \"barbican-db-sync-bhxd2\" (UID: \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\") " pod="openstack/barbican-db-sync-bhxd2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813091 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrq5k\" (UniqueName: \"kubernetes.io/projected/9c7939fb-33c3-486d-85a6-05265f29d3e8-kube-api-access-rrq5k\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813123 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-config-data\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813141 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd4s2\" (UniqueName: \"kubernetes.io/projected/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-kube-api-access-hd4s2\") pod \"barbican-db-sync-bhxd2\" (UID: \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\") " pod="openstack/barbican-db-sync-bhxd2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813165 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395cb84c-684a-48ec-aa24-03f5a7f218ff-run-httpd\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813184 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-config-data\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813204 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813245 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395cb84c-684a-48ec-aa24-03f5a7f218ff-log-httpd\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813267 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-db-sync-config-data\") pod \"barbican-db-sync-bhxd2\" (UID: \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\") " pod="openstack/barbican-db-sync-bhxd2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813300 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-scripts\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813328 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c7939fb-33c3-486d-85a6-05265f29d3e8-logs\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813349 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813375 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-combined-ca-bundle\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.813390 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-scripts\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.824678 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-db-sync-config-data\") pod \"barbican-db-sync-bhxd2\" (UID: \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\") " pod="openstack/barbican-db-sync-bhxd2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.830573 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-combined-ca-bundle\") pod \"barbican-db-sync-bhxd2\" (UID: \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\") " pod="openstack/barbican-db-sync-bhxd2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.837816 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gsljc" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.842515 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd4s2\" (UniqueName: \"kubernetes.io/projected/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-kube-api-access-hd4s2\") pod \"barbican-db-sync-bhxd2\" (UID: \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\") " pod="openstack/barbican-db-sync-bhxd2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.910427 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bhxd2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.920982 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-config\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921029 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-scripts\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921168 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c7939fb-33c3-486d-85a6-05265f29d3e8-logs\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921194 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921241 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-combined-ca-bundle\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921256 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-scripts\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921280 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjd46\" (UniqueName: \"kubernetes.io/projected/395cb84c-684a-48ec-aa24-03f5a7f218ff-kube-api-access-kjd46\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921325 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrq5k\" (UniqueName: \"kubernetes.io/projected/9c7939fb-33c3-486d-85a6-05265f29d3e8-kube-api-access-rrq5k\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921350 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-config-data\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921396 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921421 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfnbj\" (UniqueName: \"kubernetes.io/projected/b7262962-40ae-43ac-af47-bf905eb7c920-kube-api-access-hfnbj\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921437 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395cb84c-684a-48ec-aa24-03f5a7f218ff-run-httpd\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921458 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-config-data\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921478 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921496 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-dns-svc\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921538 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395cb84c-684a-48ec-aa24-03f5a7f218ff-log-httpd\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.921563 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.922702 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c7939fb-33c3-486d-85a6-05265f29d3e8-logs\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.924082 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395cb84c-684a-48ec-aa24-03f5a7f218ff-log-httpd\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.924397 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395cb84c-684a-48ec-aa24-03f5a7f218ff-run-httpd\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.927969 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.961729 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.963028 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-scripts\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.979848 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrq5k\" (UniqueName: \"kubernetes.io/projected/9c7939fb-33c3-486d-85a6-05265f29d3e8-kube-api-access-rrq5k\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:32 crc kubenswrapper[4726]: I1001 16:16:32.985263 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-scripts\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.024970 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.025230 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfnbj\" (UniqueName: \"kubernetes.io/projected/b7262962-40ae-43ac-af47-bf905eb7c920-kube-api-access-hfnbj\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.025352 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-dns-svc\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.025562 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.025624 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-config\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.034797 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-dns-svc\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.041998 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-config\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.043087 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.043912 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-config-data\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.043922 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-combined-ca-bundle\") pod \"placement-db-sync-ks2wk\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.044226 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.044704 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-config-data\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.047894 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-m7bvd"] Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.075908 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjd46\" (UniqueName: \"kubernetes.io/projected/395cb84c-684a-48ec-aa24-03f5a7f218ff-kube-api-access-kjd46\") pod \"ceilometer-0\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " pod="openstack/ceilometer-0" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.079872 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfnbj\" (UniqueName: \"kubernetes.io/projected/b7262962-40ae-43ac-af47-bf905eb7c920-kube-api-access-hfnbj\") pod \"dnsmasq-dns-6ffb94d8ff-jvrp2\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.098898 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zwwsm"] Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.245780 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.291207 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7945d7f85f-2qfqm"] Oct 01 16:16:33 crc kubenswrapper[4726]: W1001 16:16:33.302902 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2911957_529c_46a4_8b63_e98beb36809b.slice/crio-a78c36cc993a32aa321505fadae8d604c2cf99733fb6eeaacf292b46c3cf925b WatchSource:0}: Error finding container a78c36cc993a32aa321505fadae8d604c2cf99733fb6eeaacf292b46c3cf925b: Status 404 returned error can't find the container with id a78c36cc993a32aa321505fadae8d604c2cf99733fb6eeaacf292b46c3cf925b Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.326043 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.363913 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ks2wk" Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.585960 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-skmf7"] Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.596196 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-fc7f8d489-j7grp"] Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.625896 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-gsljc"] Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.748214 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-bhxd2"] Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.762331 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-jvrp2"] Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.819181 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zwwsm" event={"ID":"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec","Type":"ContainerStarted","Data":"40273eb19a052d03ad0a5505d46cd69367281f9e6183eea03228b7bdb51acebc"} Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.819219 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7945d7f85f-2qfqm" event={"ID":"a2911957-529c-46a4-8b63-e98beb36809b","Type":"ContainerStarted","Data":"a78c36cc993a32aa321505fadae8d604c2cf99733fb6eeaacf292b46c3cf925b"} Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.819230 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" event={"ID":"a8c3496b-70c5-4e68-8781-b4936a302273","Type":"ContainerStarted","Data":"1b662b8d50f1a25397c42fb24e20009854be04e20b3a7e12c7a745da20d52b84"} Oct 01 16:16:33 crc kubenswrapper[4726]: I1001 16:16:33.850090 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=71.501120268 podStartE2EDuration="1m34.850047983s" podCreationTimestamp="2025-10-01 16:14:59 +0000 UTC" firstStartedPulling="2025-10-01 16:16:07.248289417 +0000 UTC m=+1020.149841994" lastFinishedPulling="2025-10-01 16:16:30.597217142 +0000 UTC m=+1043.498769709" observedRunningTime="2025-10-01 16:16:33.844924026 +0000 UTC m=+1046.746476613" watchObservedRunningTime="2025-10-01 16:16:33.850047983 +0000 UTC m=+1046.751600560" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.104160 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-jvrp2"] Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.132507 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-f594w"] Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.138221 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.147104 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.172872 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-f594w"] Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.255882 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.255928 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-dns-svc\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.255989 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.256080 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.256108 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-config\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.256136 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrqhr\" (UniqueName: \"kubernetes.io/projected/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-kube-api-access-nrqhr\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.357410 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.357506 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.357539 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-config\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.357566 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrqhr\" (UniqueName: \"kubernetes.io/projected/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-kube-api-access-nrqhr\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.357589 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.357608 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-dns-svc\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.358300 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.358368 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-config\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.361715 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.361854 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-dns-svc\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.362122 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.377000 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrqhr\" (UniqueName: \"kubernetes.io/projected/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-kube-api-access-nrqhr\") pod \"dnsmasq-dns-cf78879c9-f594w\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.464436 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.794795 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7945d7f85f-2qfqm"] Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.832517 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-69b99cdcfc-z8t52"] Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.835457 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.851389 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-69b99cdcfc-z8t52"] Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.863814 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.968083 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4cff8692-89be-4a73-b714-1cc1e23a6b7d-scripts\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.968161 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cff8692-89be-4a73-b714-1cc1e23a6b7d-logs\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.968189 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hknfw\" (UniqueName: \"kubernetes.io/projected/4cff8692-89be-4a73-b714-1cc1e23a6b7d-kube-api-access-hknfw\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.968290 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4cff8692-89be-4a73-b714-1cc1e23a6b7d-config-data\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:34 crc kubenswrapper[4726]: I1001 16:16:34.968368 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4cff8692-89be-4a73-b714-1cc1e23a6b7d-horizon-secret-key\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:35 crc kubenswrapper[4726]: I1001 16:16:35.070129 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4cff8692-89be-4a73-b714-1cc1e23a6b7d-horizon-secret-key\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:35 crc kubenswrapper[4726]: I1001 16:16:35.070210 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4cff8692-89be-4a73-b714-1cc1e23a6b7d-scripts\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:35 crc kubenswrapper[4726]: I1001 16:16:35.070230 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cff8692-89be-4a73-b714-1cc1e23a6b7d-logs\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:35 crc kubenswrapper[4726]: I1001 16:16:35.070254 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hknfw\" (UniqueName: \"kubernetes.io/projected/4cff8692-89be-4a73-b714-1cc1e23a6b7d-kube-api-access-hknfw\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:35 crc kubenswrapper[4726]: I1001 16:16:35.070317 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4cff8692-89be-4a73-b714-1cc1e23a6b7d-config-data\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:35 crc kubenswrapper[4726]: I1001 16:16:35.071369 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4cff8692-89be-4a73-b714-1cc1e23a6b7d-config-data\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:35 crc kubenswrapper[4726]: I1001 16:16:35.072180 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cff8692-89be-4a73-b714-1cc1e23a6b7d-logs\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:35 crc kubenswrapper[4726]: I1001 16:16:35.072328 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4cff8692-89be-4a73-b714-1cc1e23a6b7d-scripts\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:35 crc kubenswrapper[4726]: I1001 16:16:35.078021 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4cff8692-89be-4a73-b714-1cc1e23a6b7d-horizon-secret-key\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:35 crc kubenswrapper[4726]: I1001 16:16:35.097204 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hknfw\" (UniqueName: \"kubernetes.io/projected/4cff8692-89be-4a73-b714-1cc1e23a6b7d-kube-api-access-hknfw\") pod \"horizon-69b99cdcfc-z8t52\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:35 crc kubenswrapper[4726]: I1001 16:16:35.151322 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:16:36 crc kubenswrapper[4726]: W1001 16:16:36.043720 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b994ac8_f28c_458a_873f_4270e9f0f36b.slice/crio-7d3c1db43ffc80652e318211f6b2db23b9f65f3c63eae25c624bb98a6a8f3a8b WatchSource:0}: Error finding container 7d3c1db43ffc80652e318211f6b2db23b9f65f3c63eae25c624bb98a6a8f3a8b: Status 404 returned error can't find the container with id 7d3c1db43ffc80652e318211f6b2db23b9f65f3c63eae25c624bb98a6a8f3a8b Oct 01 16:16:36 crc kubenswrapper[4726]: W1001 16:16:36.046761 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62a37356_6f11_4c75_a452_570a847a0518.slice/crio-27393fa6ac70e4f5c2bb8f970c76f5d94f054021da3e82838b4ec2c2798e7363 WatchSource:0}: Error finding container 27393fa6ac70e4f5c2bb8f970c76f5d94f054021da3e82838b4ec2c2798e7363: Status 404 returned error can't find the container with id 27393fa6ac70e4f5c2bb8f970c76f5d94f054021da3e82838b4ec2c2798e7363 Oct 01 16:16:36 crc kubenswrapper[4726]: W1001 16:16:36.058025 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda83273b9_e1f0_47e6_8bbe_839f4c0075b3.slice/crio-bb7c90e518c145514f1782850e7d523b3c628c6879372caf741bd9a193272360 WatchSource:0}: Error finding container bb7c90e518c145514f1782850e7d523b3c628c6879372caf741bd9a193272360: Status 404 returned error can't find the container with id bb7c90e518c145514f1782850e7d523b3c628c6879372caf741bd9a193272360 Oct 01 16:16:36 crc kubenswrapper[4726]: W1001 16:16:36.062487 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdb5fc73_43ce_425a_8a5f_cd1b3b96aba6.slice/crio-f57d82a6060294c1b43e8117ffc939abf7245e8f1ff27b262bcff2a4e8093ecb WatchSource:0}: Error finding container f57d82a6060294c1b43e8117ffc939abf7245e8f1ff27b262bcff2a4e8093ecb: Status 404 returned error can't find the container with id f57d82a6060294c1b43e8117ffc939abf7245e8f1ff27b262bcff2a4e8093ecb Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.583958 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-69b99cdcfc-z8t52"] Oct 01 16:16:36 crc kubenswrapper[4726]: W1001 16:16:36.592038 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4cff8692_89be_4a73_b714_1cc1e23a6b7d.slice/crio-e36f6ec5d434f48d3fc32b864c1d3c45c7a2d4e6614a5b7e2158dca1a2417af8 WatchSource:0}: Error finding container e36f6ec5d434f48d3fc32b864c1d3c45c7a2d4e6614a5b7e2158dca1a2417af8: Status 404 returned error can't find the container with id e36f6ec5d434f48d3fc32b864c1d3c45c7a2d4e6614a5b7e2158dca1a2417af8 Oct 01 16:16:36 crc kubenswrapper[4726]: W1001 16:16:36.686253 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod395cb84c_684a_48ec_aa24_03f5a7f218ff.slice/crio-9fef1c6274d546d30da9bf9a726cd36c9820448482b452f0163a55540d88fb55 WatchSource:0}: Error finding container 9fef1c6274d546d30da9bf9a726cd36c9820448482b452f0163a55540d88fb55: Status 404 returned error can't find the container with id 9fef1c6274d546d30da9bf9a726cd36c9820448482b452f0163a55540d88fb55 Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.686868 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.732088 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ks2wk"] Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.750490 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-f594w"] Oct 01 16:16:36 crc kubenswrapper[4726]: W1001 16:16:36.756177 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5c407e4_2f8c_47dd_85b4_b29c12ea3aed.slice/crio-9c7797c6d4e7f2728b0d1cae24942ed994e8ae747c67184276ca993b91c9b787 WatchSource:0}: Error finding container 9c7797c6d4e7f2728b0d1cae24942ed994e8ae747c67184276ca993b91c9b787: Status 404 returned error can't find the container with id 9c7797c6d4e7f2728b0d1cae24942ed994e8ae747c67184276ca993b91c9b787 Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.866150 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-skmf7" event={"ID":"7b994ac8-f28c-458a-873f-4270e9f0f36b","Type":"ContainerStarted","Data":"7d3c1db43ffc80652e318211f6b2db23b9f65f3c63eae25c624bb98a6a8f3a8b"} Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.867675 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bhxd2" event={"ID":"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6","Type":"ContainerStarted","Data":"f57d82a6060294c1b43e8117ffc939abf7245e8f1ff27b262bcff2a4e8093ecb"} Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.869120 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69b99cdcfc-z8t52" event={"ID":"4cff8692-89be-4a73-b714-1cc1e23a6b7d","Type":"ContainerStarted","Data":"e36f6ec5d434f48d3fc32b864c1d3c45c7a2d4e6614a5b7e2158dca1a2417af8"} Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.870408 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" event={"ID":"a8c3496b-70c5-4e68-8781-b4936a302273","Type":"ContainerStarted","Data":"d5afbe45fade0c968d90fba1ebf0fbaa8e05addc4ccfe0d676f840202b4125e3"} Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.872887 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" event={"ID":"b7262962-40ae-43ac-af47-bf905eb7c920","Type":"ContainerStarted","Data":"694ea2f74fca26a17221f120fe7a7ec3a359894a18b15977d01479536f6719bf"} Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.874361 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fc7f8d489-j7grp" event={"ID":"62a37356-6f11-4c75-a452-570a847a0518","Type":"ContainerStarted","Data":"27393fa6ac70e4f5c2bb8f970c76f5d94f054021da3e82838b4ec2c2798e7363"} Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.875593 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ks2wk" event={"ID":"9c7939fb-33c3-486d-85a6-05265f29d3e8","Type":"ContainerStarted","Data":"484bfc6ee63f2859d77f581ea837c8101abe4689fb5d35d30836f046c711ebd4"} Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.878684 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395cb84c-684a-48ec-aa24-03f5a7f218ff","Type":"ContainerStarted","Data":"9fef1c6274d546d30da9bf9a726cd36c9820448482b452f0163a55540d88fb55"} Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.879825 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gsljc" event={"ID":"a83273b9-e1f0-47e6-8bbe-839f4c0075b3","Type":"ContainerStarted","Data":"bb7c90e518c145514f1782850e7d523b3c628c6879372caf741bd9a193272360"} Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.881625 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-f594w" event={"ID":"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed","Type":"ContainerStarted","Data":"9c7797c6d4e7f2728b0d1cae24942ed994e8ae747c67184276ca993b91c9b787"} Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.883742 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zwwsm" event={"ID":"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec","Type":"ContainerStarted","Data":"7d6aecbe07d7f7d01890bd04f0856e0f85bde44f6889434a4b4d2b78c0feed64"} Oct 01 16:16:36 crc kubenswrapper[4726]: I1001 16:16:36.906176 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-zwwsm" podStartSLOduration=4.906160954 podStartE2EDuration="4.906160954s" podCreationTimestamp="2025-10-01 16:16:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:16:36.901864991 +0000 UTC m=+1049.803417578" watchObservedRunningTime="2025-10-01 16:16:36.906160954 +0000 UTC m=+1049.807713541" Oct 01 16:16:37 crc kubenswrapper[4726]: I1001 16:16:37.900207 4726 generic.go:334] "Generic (PLEG): container finished" podID="b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" containerID="7afdceed352942813bfe7a35b00f27efbf39f21902a199081e5e2deaabce7ee4" exitCode=0 Oct 01 16:16:37 crc kubenswrapper[4726]: I1001 16:16:37.900743 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-f594w" event={"ID":"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed","Type":"ContainerDied","Data":"7afdceed352942813bfe7a35b00f27efbf39f21902a199081e5e2deaabce7ee4"} Oct 01 16:16:37 crc kubenswrapper[4726]: I1001 16:16:37.910187 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8c3496b-70c5-4e68-8781-b4936a302273" containerID="d5afbe45fade0c968d90fba1ebf0fbaa8e05addc4ccfe0d676f840202b4125e3" exitCode=0 Oct 01 16:16:37 crc kubenswrapper[4726]: I1001 16:16:37.910270 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" event={"ID":"a8c3496b-70c5-4e68-8781-b4936a302273","Type":"ContainerDied","Data":"d5afbe45fade0c968d90fba1ebf0fbaa8e05addc4ccfe0d676f840202b4125e3"} Oct 01 16:16:37 crc kubenswrapper[4726]: I1001 16:16:37.929698 4726 generic.go:334] "Generic (PLEG): container finished" podID="b7262962-40ae-43ac-af47-bf905eb7c920" containerID="5e7b389f3a71a1a7ddd01bb9c1871e5973f18162796c0f52c65384f3cfd882f4" exitCode=0 Oct 01 16:16:37 crc kubenswrapper[4726]: I1001 16:16:37.929986 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" event={"ID":"b7262962-40ae-43ac-af47-bf905eb7c920","Type":"ContainerDied","Data":"5e7b389f3a71a1a7ddd01bb9c1871e5973f18162796c0f52c65384f3cfd882f4"} Oct 01 16:16:37 crc kubenswrapper[4726]: I1001 16:16:37.941223 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-skmf7" event={"ID":"7b994ac8-f28c-458a-873f-4270e9f0f36b","Type":"ContainerStarted","Data":"79416f7542f476fdc76e78a15e310f7390f6670c401eb02b3b186018905708d4"} Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.265169 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-skmf7" podStartSLOduration=6.2651475340000005 podStartE2EDuration="6.265147534s" podCreationTimestamp="2025-10-01 16:16:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:16:38.240990815 +0000 UTC m=+1051.142543392" watchObservedRunningTime="2025-10-01 16:16:38.265147534 +0000 UTC m=+1051.166700111" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.419566 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.580470 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.600385 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-dns-svc\") pod \"b7262962-40ae-43ac-af47-bf905eb7c920\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.600448 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-ovsdbserver-nb\") pod \"b7262962-40ae-43ac-af47-bf905eb7c920\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.600567 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfnbj\" (UniqueName: \"kubernetes.io/projected/b7262962-40ae-43ac-af47-bf905eb7c920-kube-api-access-hfnbj\") pod \"b7262962-40ae-43ac-af47-bf905eb7c920\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.600664 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-config\") pod \"b7262962-40ae-43ac-af47-bf905eb7c920\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.600769 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-ovsdbserver-sb\") pod \"b7262962-40ae-43ac-af47-bf905eb7c920\" (UID: \"b7262962-40ae-43ac-af47-bf905eb7c920\") " Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.617339 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7262962-40ae-43ac-af47-bf905eb7c920-kube-api-access-hfnbj" (OuterVolumeSpecName: "kube-api-access-hfnbj") pod "b7262962-40ae-43ac-af47-bf905eb7c920" (UID: "b7262962-40ae-43ac-af47-bf905eb7c920"). InnerVolumeSpecName "kube-api-access-hfnbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.631626 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b7262962-40ae-43ac-af47-bf905eb7c920" (UID: "b7262962-40ae-43ac-af47-bf905eb7c920"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.642241 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b7262962-40ae-43ac-af47-bf905eb7c920" (UID: "b7262962-40ae-43ac-af47-bf905eb7c920"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.642721 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b7262962-40ae-43ac-af47-bf905eb7c920" (UID: "b7262962-40ae-43ac-af47-bf905eb7c920"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.653141 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-config" (OuterVolumeSpecName: "config") pod "b7262962-40ae-43ac-af47-bf905eb7c920" (UID: "b7262962-40ae-43ac-af47-bf905eb7c920"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.705949 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-ovsdbserver-nb\") pod \"a8c3496b-70c5-4e68-8781-b4936a302273\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.706016 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2tqk\" (UniqueName: \"kubernetes.io/projected/a8c3496b-70c5-4e68-8781-b4936a302273-kube-api-access-d2tqk\") pod \"a8c3496b-70c5-4e68-8781-b4936a302273\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.706103 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-config\") pod \"a8c3496b-70c5-4e68-8781-b4936a302273\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.706147 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-ovsdbserver-sb\") pod \"a8c3496b-70c5-4e68-8781-b4936a302273\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.706167 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-dns-svc\") pod \"a8c3496b-70c5-4e68-8781-b4936a302273\" (UID: \"a8c3496b-70c5-4e68-8781-b4936a302273\") " Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.708157 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.708193 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.708206 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.708217 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfnbj\" (UniqueName: \"kubernetes.io/projected/b7262962-40ae-43ac-af47-bf905eb7c920-kube-api-access-hfnbj\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.708235 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7262962-40ae-43ac-af47-bf905eb7c920-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.711700 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8c3496b-70c5-4e68-8781-b4936a302273-kube-api-access-d2tqk" (OuterVolumeSpecName: "kube-api-access-d2tqk") pod "a8c3496b-70c5-4e68-8781-b4936a302273" (UID: "a8c3496b-70c5-4e68-8781-b4936a302273"). InnerVolumeSpecName "kube-api-access-d2tqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.731479 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-config" (OuterVolumeSpecName: "config") pod "a8c3496b-70c5-4e68-8781-b4936a302273" (UID: "a8c3496b-70c5-4e68-8781-b4936a302273"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.744758 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a8c3496b-70c5-4e68-8781-b4936a302273" (UID: "a8c3496b-70c5-4e68-8781-b4936a302273"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.751711 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a8c3496b-70c5-4e68-8781-b4936a302273" (UID: "a8c3496b-70c5-4e68-8781-b4936a302273"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.756463 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a8c3496b-70c5-4e68-8781-b4936a302273" (UID: "a8c3496b-70c5-4e68-8781-b4936a302273"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.812199 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.812559 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2tqk\" (UniqueName: \"kubernetes.io/projected/a8c3496b-70c5-4e68-8781-b4936a302273-kube-api-access-d2tqk\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.812587 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.812597 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.812606 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8c3496b-70c5-4e68-8781-b4936a302273-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.972216 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-f594w" event={"ID":"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed","Type":"ContainerStarted","Data":"aee4c133a88a3f9da53ef17d7867178dda83f566fe8fa055284ea1039ca52a6f"} Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.972292 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.991913 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" event={"ID":"a8c3496b-70c5-4e68-8781-b4936a302273","Type":"ContainerDied","Data":"1b662b8d50f1a25397c42fb24e20009854be04e20b3a7e12c7a745da20d52b84"} Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.991963 4726 scope.go:117] "RemoveContainer" containerID="d5afbe45fade0c968d90fba1ebf0fbaa8e05addc4ccfe0d676f840202b4125e3" Oct 01 16:16:38 crc kubenswrapper[4726]: I1001 16:16:38.992085 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-m7bvd" Oct 01 16:16:39 crc kubenswrapper[4726]: I1001 16:16:39.001084 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" Oct 01 16:16:39 crc kubenswrapper[4726]: I1001 16:16:39.001606 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-jvrp2" event={"ID":"b7262962-40ae-43ac-af47-bf905eb7c920","Type":"ContainerDied","Data":"694ea2f74fca26a17221f120fe7a7ec3a359894a18b15977d01479536f6719bf"} Oct 01 16:16:39 crc kubenswrapper[4726]: I1001 16:16:39.007565 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf78879c9-f594w" podStartSLOduration=5.007545752 podStartE2EDuration="5.007545752s" podCreationTimestamp="2025-10-01 16:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:16:38.99978215 +0000 UTC m=+1051.901334727" watchObservedRunningTime="2025-10-01 16:16:39.007545752 +0000 UTC m=+1051.909098329" Oct 01 16:16:39 crc kubenswrapper[4726]: I1001 16:16:39.031907 4726 scope.go:117] "RemoveContainer" containerID="5e7b389f3a71a1a7ddd01bb9c1871e5973f18162796c0f52c65384f3cfd882f4" Oct 01 16:16:39 crc kubenswrapper[4726]: I1001 16:16:39.053943 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-m7bvd"] Oct 01 16:16:39 crc kubenswrapper[4726]: I1001 16:16:39.072135 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-m7bvd"] Oct 01 16:16:39 crc kubenswrapper[4726]: I1001 16:16:39.104190 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-jvrp2"] Oct 01 16:16:39 crc kubenswrapper[4726]: I1001 16:16:39.111687 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-jvrp2"] Oct 01 16:16:39 crc kubenswrapper[4726]: I1001 16:16:39.818154 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8c3496b-70c5-4e68-8781-b4936a302273" path="/var/lib/kubelet/pods/a8c3496b-70c5-4e68-8781-b4936a302273/volumes" Oct 01 16:16:39 crc kubenswrapper[4726]: I1001 16:16:39.819149 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7262962-40ae-43ac-af47-bf905eb7c920" path="/var/lib/kubelet/pods/b7262962-40ae-43ac-af47-bf905eb7c920/volumes" Oct 01 16:16:40 crc kubenswrapper[4726]: I1001 16:16:40.013745 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zwfrl" event={"ID":"bdf2f47d-bf60-4b49-ae78-1b7823838797","Type":"ContainerStarted","Data":"5d95995d161ada2dffe5ebe08d7905b3a6343395a2fa4d121ffbfa5dd587f3c5"} Oct 01 16:16:40 crc kubenswrapper[4726]: I1001 16:16:40.032464 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-zwfrl" podStartSLOduration=3.202549244 podStartE2EDuration="43.032448329s" podCreationTimestamp="2025-10-01 16:15:57 +0000 UTC" firstStartedPulling="2025-10-01 16:15:58.826491725 +0000 UTC m=+1011.728044302" lastFinishedPulling="2025-10-01 16:16:38.65639081 +0000 UTC m=+1051.557943387" observedRunningTime="2025-10-01 16:16:40.027389815 +0000 UTC m=+1052.928942392" watchObservedRunningTime="2025-10-01 16:16:40.032448329 +0000 UTC m=+1052.934000906" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.623889 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-fc7f8d489-j7grp"] Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.698653 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c6ccb6664-v52zf"] Oct 01 16:16:41 crc kubenswrapper[4726]: E1001 16:16:41.699878 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8c3496b-70c5-4e68-8781-b4936a302273" containerName="init" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.699990 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8c3496b-70c5-4e68-8781-b4936a302273" containerName="init" Oct 01 16:16:41 crc kubenswrapper[4726]: E1001 16:16:41.700106 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7262962-40ae-43ac-af47-bf905eb7c920" containerName="init" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.700184 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7262962-40ae-43ac-af47-bf905eb7c920" containerName="init" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.700463 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8c3496b-70c5-4e68-8781-b4936a302273" containerName="init" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.700556 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7262962-40ae-43ac-af47-bf905eb7c920" containerName="init" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.701803 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.719813 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c6ccb6664-v52zf"] Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.720325 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.781940 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b0861ea-bb92-4592-9129-f9b3b4725a51-config-data\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.782689 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-horizon-secret-key\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.782760 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b0861ea-bb92-4592-9129-f9b3b4725a51-logs\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.782882 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b0861ea-bb92-4592-9129-f9b3b4725a51-scripts\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.782907 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-horizon-tls-certs\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.782977 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-combined-ca-bundle\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.783006 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztxl5\" (UniqueName: \"kubernetes.io/projected/5b0861ea-bb92-4592-9129-f9b3b4725a51-kube-api-access-ztxl5\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.886159 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztxl5\" (UniqueName: \"kubernetes.io/projected/5b0861ea-bb92-4592-9129-f9b3b4725a51-kube-api-access-ztxl5\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.886258 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b0861ea-bb92-4592-9129-f9b3b4725a51-config-data\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.886289 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-horizon-secret-key\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.886331 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b0861ea-bb92-4592-9129-f9b3b4725a51-logs\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.886353 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b0861ea-bb92-4592-9129-f9b3b4725a51-scripts\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.886375 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-horizon-tls-certs\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.886415 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-combined-ca-bundle\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.895549 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b0861ea-bb92-4592-9129-f9b3b4725a51-logs\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.896601 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b0861ea-bb92-4592-9129-f9b3b4725a51-scripts\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.897096 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b0861ea-bb92-4592-9129-f9b3b4725a51-config-data\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.900757 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-69b99cdcfc-z8t52"] Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.900796 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7b5b89f8d4-d8p9g"] Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.905774 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.917805 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztxl5\" (UniqueName: \"kubernetes.io/projected/5b0861ea-bb92-4592-9129-f9b3b4725a51-kube-api-access-ztxl5\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.927946 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-combined-ca-bundle\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.930758 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-horizon-secret-key\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.930812 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b5b89f8d4-d8p9g"] Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.956574 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-horizon-tls-certs\") pod \"horizon-c6ccb6664-v52zf\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.988529 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8ee9afe-8056-4de3-bbff-524420e36290-horizon-tls-certs\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.988595 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8ee9afe-8056-4de3-bbff-524420e36290-scripts\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.988646 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8ee9afe-8056-4de3-bbff-524420e36290-logs\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.988691 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ee9afe-8056-4de3-bbff-524420e36290-combined-ca-bundle\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.988734 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c8ee9afe-8056-4de3-bbff-524420e36290-horizon-secret-key\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.988782 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8ee9afe-8056-4de3-bbff-524420e36290-config-data\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:41 crc kubenswrapper[4726]: I1001 16:16:41.988815 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ccwj\" (UniqueName: \"kubernetes.io/projected/c8ee9afe-8056-4de3-bbff-524420e36290-kube-api-access-9ccwj\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.067694 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.091178 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8ee9afe-8056-4de3-bbff-524420e36290-logs\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.091256 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ee9afe-8056-4de3-bbff-524420e36290-combined-ca-bundle\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.091295 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c8ee9afe-8056-4de3-bbff-524420e36290-horizon-secret-key\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.091340 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8ee9afe-8056-4de3-bbff-524420e36290-config-data\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.091371 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ccwj\" (UniqueName: \"kubernetes.io/projected/c8ee9afe-8056-4de3-bbff-524420e36290-kube-api-access-9ccwj\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.091422 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8ee9afe-8056-4de3-bbff-524420e36290-horizon-tls-certs\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.091446 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8ee9afe-8056-4de3-bbff-524420e36290-scripts\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.092491 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8ee9afe-8056-4de3-bbff-524420e36290-scripts\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.093652 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8ee9afe-8056-4de3-bbff-524420e36290-config-data\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.096828 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c8ee9afe-8056-4de3-bbff-524420e36290-horizon-secret-key\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.097658 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8ee9afe-8056-4de3-bbff-524420e36290-horizon-tls-certs\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.105911 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ee9afe-8056-4de3-bbff-524420e36290-combined-ca-bundle\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.109044 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8ee9afe-8056-4de3-bbff-524420e36290-logs\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.109841 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ccwj\" (UniqueName: \"kubernetes.io/projected/c8ee9afe-8056-4de3-bbff-524420e36290-kube-api-access-9ccwj\") pod \"horizon-7b5b89f8d4-d8p9g\" (UID: \"c8ee9afe-8056-4de3-bbff-524420e36290\") " pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:42 crc kubenswrapper[4726]: I1001 16:16:42.320848 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:16:43 crc kubenswrapper[4726]: E1001 16:16:43.470695 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c9b63c2_5f33_4109_b4f7_9bf8733f65ec.slice/crio-conmon-7d6aecbe07d7f7d01890bd04f0856e0f85bde44f6889434a4b4d2b78c0feed64.scope\": RecentStats: unable to find data in memory cache]" Oct 01 16:16:44 crc kubenswrapper[4726]: I1001 16:16:44.066860 4726 generic.go:334] "Generic (PLEG): container finished" podID="1c9b63c2-5f33-4109-b4f7-9bf8733f65ec" containerID="7d6aecbe07d7f7d01890bd04f0856e0f85bde44f6889434a4b4d2b78c0feed64" exitCode=0 Oct 01 16:16:44 crc kubenswrapper[4726]: I1001 16:16:44.066907 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zwwsm" event={"ID":"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec","Type":"ContainerDied","Data":"7d6aecbe07d7f7d01890bd04f0856e0f85bde44f6889434a4b4d2b78c0feed64"} Oct 01 16:16:44 crc kubenswrapper[4726]: I1001 16:16:44.466360 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:16:44 crc kubenswrapper[4726]: I1001 16:16:44.535155 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-4rwkl"] Oct 01 16:16:44 crc kubenswrapper[4726]: I1001 16:16:44.535440 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" podUID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" containerName="dnsmasq-dns" containerID="cri-o://d4f83b310175d8a9f1e888079fe8cd4c37ea7f6bf516e002be0c8502f604e531" gracePeriod=10 Oct 01 16:16:44 crc kubenswrapper[4726]: I1001 16:16:44.651333 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" podUID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: connect: connection refused" Oct 01 16:16:45 crc kubenswrapper[4726]: I1001 16:16:45.077722 4726 generic.go:334] "Generic (PLEG): container finished" podID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" containerID="d4f83b310175d8a9f1e888079fe8cd4c37ea7f6bf516e002be0c8502f604e531" exitCode=0 Oct 01 16:16:45 crc kubenswrapper[4726]: I1001 16:16:45.077786 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" event={"ID":"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3","Type":"ContainerDied","Data":"d4f83b310175d8a9f1e888079fe8cd4c37ea7f6bf516e002be0c8502f604e531"} Oct 01 16:16:49 crc kubenswrapper[4726]: I1001 16:16:49.651329 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" podUID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: connect: connection refused" Oct 01 16:16:53 crc kubenswrapper[4726]: E1001 16:16:53.910664 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Oct 01 16:16:53 crc kubenswrapper[4726]: E1001 16:16:53.911394 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n657h78hc8h5c7h64bh547h678h645h565h66h69h68ch59bh59fhfdh56dh85h57ch57bhfh5cchfdh548h5h54dhf7h5c6hf7h5c4hb5h586hf6q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kjd46,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(395cb84c-684a-48ec-aa24-03f5a7f218ff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:16:53 crc kubenswrapper[4726]: I1001 16:16:53.984186 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.136116 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-scripts\") pod \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.136611 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w58w\" (UniqueName: \"kubernetes.io/projected/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-kube-api-access-7w58w\") pod \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.136645 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-config-data\") pod \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.136722 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-credential-keys\") pod \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.136761 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-combined-ca-bundle\") pod \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.136818 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-fernet-keys\") pod \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\" (UID: \"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec\") " Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.143034 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1c9b63c2-5f33-4109-b4f7-9bf8733f65ec" (UID: "1c9b63c2-5f33-4109-b4f7-9bf8733f65ec"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.145238 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-kube-api-access-7w58w" (OuterVolumeSpecName: "kube-api-access-7w58w") pod "1c9b63c2-5f33-4109-b4f7-9bf8733f65ec" (UID: "1c9b63c2-5f33-4109-b4f7-9bf8733f65ec"). InnerVolumeSpecName "kube-api-access-7w58w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.153426 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-scripts" (OuterVolumeSpecName: "scripts") pod "1c9b63c2-5f33-4109-b4f7-9bf8733f65ec" (UID: "1c9b63c2-5f33-4109-b4f7-9bf8733f65ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.155608 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1c9b63c2-5f33-4109-b4f7-9bf8733f65ec" (UID: "1c9b63c2-5f33-4109-b4f7-9bf8733f65ec"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.164715 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1c9b63c2-5f33-4109-b4f7-9bf8733f65ec" (UID: "1c9b63c2-5f33-4109-b4f7-9bf8733f65ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.169960 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-config-data" (OuterVolumeSpecName: "config-data") pod "1c9b63c2-5f33-4109-b4f7-9bf8733f65ec" (UID: "1c9b63c2-5f33-4109-b4f7-9bf8733f65ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.175736 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zwwsm" event={"ID":"1c9b63c2-5f33-4109-b4f7-9bf8733f65ec","Type":"ContainerDied","Data":"40273eb19a052d03ad0a5505d46cd69367281f9e6183eea03228b7bdb51acebc"} Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.175780 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40273eb19a052d03ad0a5505d46cd69367281f9e6183eea03228b7bdb51acebc" Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.175842 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zwwsm" Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.239623 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.239668 4726 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.239687 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.239705 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w58w\" (UniqueName: \"kubernetes.io/projected/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-kube-api-access-7w58w\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.239722 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:54 crc kubenswrapper[4726]: I1001 16:16:54.239738 4726 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.154181 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-zwwsm"] Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.161068 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-zwwsm"] Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.263420 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-zfnx8"] Oct 01 16:16:55 crc kubenswrapper[4726]: E1001 16:16:55.263813 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9b63c2-5f33-4109-b4f7-9bf8733f65ec" containerName="keystone-bootstrap" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.263829 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9b63c2-5f33-4109-b4f7-9bf8733f65ec" containerName="keystone-bootstrap" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.264111 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c9b63c2-5f33-4109-b4f7-9bf8733f65ec" containerName="keystone-bootstrap" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.264703 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.267164 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.267927 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.268129 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5q5v6" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.268240 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.281326 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zfnx8"] Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.368633 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-config-data\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.368737 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89pjw\" (UniqueName: \"kubernetes.io/projected/df1f486b-c5af-47de-9af6-4c499af830db-kube-api-access-89pjw\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.368887 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-fernet-keys\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.368975 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-combined-ca-bundle\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.369129 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-scripts\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.369339 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-credential-keys\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.470397 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-combined-ca-bundle\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.471235 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-scripts\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.471296 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-credential-keys\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.471376 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89pjw\" (UniqueName: \"kubernetes.io/projected/df1f486b-c5af-47de-9af6-4c499af830db-kube-api-access-89pjw\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.471394 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-config-data\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.471439 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-fernet-keys\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.475807 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-combined-ca-bundle\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.477135 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-credential-keys\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.477428 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-config-data\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.477926 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-fernet-keys\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.484445 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-scripts\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.490755 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89pjw\" (UniqueName: \"kubernetes.io/projected/df1f486b-c5af-47de-9af6-4c499af830db-kube-api-access-89pjw\") pod \"keystone-bootstrap-zfnx8\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.587140 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:16:55 crc kubenswrapper[4726]: I1001 16:16:55.818918 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c9b63c2-5f33-4109-b4f7-9bf8733f65ec" path="/var/lib/kubelet/pods/1c9b63c2-5f33-4109-b4f7-9bf8733f65ec/volumes" Oct 01 16:16:59 crc kubenswrapper[4726]: I1001 16:16:59.651018 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" podUID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: i/o timeout" Oct 01 16:16:59 crc kubenswrapper[4726]: I1001 16:16:59.651654 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:17:01 crc kubenswrapper[4726]: I1001 16:17:01.238366 4726 generic.go:334] "Generic (PLEG): container finished" podID="bdf2f47d-bf60-4b49-ae78-1b7823838797" containerID="5d95995d161ada2dffe5ebe08d7905b3a6343395a2fa4d121ffbfa5dd587f3c5" exitCode=0 Oct 01 16:17:01 crc kubenswrapper[4726]: I1001 16:17:01.238526 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zwfrl" event={"ID":"bdf2f47d-bf60-4b49-ae78-1b7823838797","Type":"ContainerDied","Data":"5d95995d161ada2dffe5ebe08d7905b3a6343395a2fa4d121ffbfa5dd587f3c5"} Oct 01 16:17:04 crc kubenswrapper[4726]: I1001 16:17:04.651860 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" podUID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: i/o timeout" Oct 01 16:17:09 crc kubenswrapper[4726]: I1001 16:17:09.652759 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" podUID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: i/o timeout" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.240673 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.276854 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zwfrl" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.375268 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zwfrl" event={"ID":"bdf2f47d-bf60-4b49-ae78-1b7823838797","Type":"ContainerDied","Data":"69d68905f8bc2dfd35245d7fb80d3858575b8d612170ff741c6a27183cb3f9e2"} Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.375592 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69d68905f8bc2dfd35245d7fb80d3858575b8d612170ff741c6a27183cb3f9e2" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.375451 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zwfrl" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.377443 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" event={"ID":"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3","Type":"ContainerDied","Data":"be518624d44f984cce7f96bd4920ea8189151cbf9eed04ac82221e209a28e67f"} Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.377504 4726 scope.go:117] "RemoveContainer" containerID="d4f83b310175d8a9f1e888079fe8cd4c37ea7f6bf516e002be0c8502f604e531" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.377515 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.382638 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-combined-ca-bundle\") pod \"bdf2f47d-bf60-4b49-ae78-1b7823838797\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.382725 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-db-sync-config-data\") pod \"bdf2f47d-bf60-4b49-ae78-1b7823838797\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.382783 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-dns-svc\") pod \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.382878 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-ovsdbserver-sb\") pod \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.382905 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-config-data\") pod \"bdf2f47d-bf60-4b49-ae78-1b7823838797\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.382980 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-ovsdbserver-nb\") pod \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.383011 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-config\") pod \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.383104 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ztmm\" (UniqueName: \"kubernetes.io/projected/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-kube-api-access-6ztmm\") pod \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\" (UID: \"e34367e7-1d09-41ef-ad55-e11ce0e5f2a3\") " Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.383158 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7pjv\" (UniqueName: \"kubernetes.io/projected/bdf2f47d-bf60-4b49-ae78-1b7823838797-kube-api-access-f7pjv\") pod \"bdf2f47d-bf60-4b49-ae78-1b7823838797\" (UID: \"bdf2f47d-bf60-4b49-ae78-1b7823838797\") " Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.395018 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bdf2f47d-bf60-4b49-ae78-1b7823838797" (UID: "bdf2f47d-bf60-4b49-ae78-1b7823838797"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.396934 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-kube-api-access-6ztmm" (OuterVolumeSpecName: "kube-api-access-6ztmm") pod "e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" (UID: "e34367e7-1d09-41ef-ad55-e11ce0e5f2a3"). InnerVolumeSpecName "kube-api-access-6ztmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.397012 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdf2f47d-bf60-4b49-ae78-1b7823838797-kube-api-access-f7pjv" (OuterVolumeSpecName: "kube-api-access-f7pjv") pod "bdf2f47d-bf60-4b49-ae78-1b7823838797" (UID: "bdf2f47d-bf60-4b49-ae78-1b7823838797"). InnerVolumeSpecName "kube-api-access-f7pjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.438488 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdf2f47d-bf60-4b49-ae78-1b7823838797" (UID: "bdf2f47d-bf60-4b49-ae78-1b7823838797"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.441103 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" (UID: "e34367e7-1d09-41ef-ad55-e11ce0e5f2a3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.441556 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" (UID: "e34367e7-1d09-41ef-ad55-e11ce0e5f2a3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.446289 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-config-data" (OuterVolumeSpecName: "config-data") pod "bdf2f47d-bf60-4b49-ae78-1b7823838797" (UID: "bdf2f47d-bf60-4b49-ae78-1b7823838797"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.447371 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" (UID: "e34367e7-1d09-41ef-ad55-e11ce0e5f2a3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.448136 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-config" (OuterVolumeSpecName: "config") pod "e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" (UID: "e34367e7-1d09-41ef-ad55-e11ce0e5f2a3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:14 crc kubenswrapper[4726]: E1001 16:17:14.452253 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 01 16:17:14 crc kubenswrapper[4726]: E1001 16:17:14.452402 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hd4s2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-bhxd2_openstack(fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:17:14 crc kubenswrapper[4726]: E1001 16:17:14.454255 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-bhxd2" podUID="fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.485144 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.485220 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.485232 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ztmm\" (UniqueName: \"kubernetes.io/projected/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-kube-api-access-6ztmm\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.485243 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7pjv\" (UniqueName: \"kubernetes.io/projected/bdf2f47d-bf60-4b49-ae78-1b7823838797-kube-api-access-f7pjv\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.485252 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.485283 4726 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.485293 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.485301 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.485310 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf2f47d-bf60-4b49-ae78-1b7823838797-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.549128 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b5b89f8d4-d8p9g"] Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.654494 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-4rwkl" podUID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: i/o timeout" Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.731732 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-4rwkl"] Oct 01 16:17:14 crc kubenswrapper[4726]: I1001 16:17:14.741859 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-4rwkl"] Oct 01 16:17:15 crc kubenswrapper[4726]: E1001 16:17:15.390635 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-bhxd2" podUID="fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6" Oct 01 16:17:15 crc kubenswrapper[4726]: E1001 16:17:15.623486 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 01 16:17:15 crc kubenswrapper[4726]: E1001 16:17:15.623874 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-znxm2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-gsljc_openstack(a83273b9-e1f0-47e6-8bbe-839f4c0075b3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:17:15 crc kubenswrapper[4726]: E1001 16:17:15.625299 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-gsljc" podUID="a83273b9-e1f0-47e6-8bbe-839f4c0075b3" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.655355 4726 scope.go:117] "RemoveContainer" containerID="3b75a4eec0509e57bec3226a4fcd9a24e96dfba115a8e09a91fb952d564df1c2" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.770228 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bdn6t"] Oct 01 16:17:15 crc kubenswrapper[4726]: E1001 16:17:15.770711 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" containerName="init" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.770723 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" containerName="init" Oct 01 16:17:15 crc kubenswrapper[4726]: E1001 16:17:15.770768 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" containerName="dnsmasq-dns" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.770775 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" containerName="dnsmasq-dns" Oct 01 16:17:15 crc kubenswrapper[4726]: E1001 16:17:15.770792 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf2f47d-bf60-4b49-ae78-1b7823838797" containerName="glance-db-sync" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.770798 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf2f47d-bf60-4b49-ae78-1b7823838797" containerName="glance-db-sync" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.771030 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf2f47d-bf60-4b49-ae78-1b7823838797" containerName="glance-db-sync" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.771124 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" containerName="dnsmasq-dns" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.772262 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.787695 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bdn6t"] Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.827372 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e34367e7-1d09-41ef-ad55-e11ce0e5f2a3" path="/var/lib/kubelet/pods/e34367e7-1d09-41ef-ad55-e11ce0e5f2a3/volumes" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.912133 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.912279 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.912306 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zqpd\" (UniqueName: \"kubernetes.io/projected/34e6cb54-5cda-4ac4-bfe1-e59986daa349-kube-api-access-6zqpd\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.912328 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-config\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.912351 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:15 crc kubenswrapper[4726]: I1001 16:17:15.912388 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.014161 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.014269 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.015253 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.015278 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.015416 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.015439 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zqpd\" (UniqueName: \"kubernetes.io/projected/34e6cb54-5cda-4ac4-bfe1-e59986daa349-kube-api-access-6zqpd\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.015962 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.015993 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-config\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.016019 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.016432 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-config\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.016727 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.031409 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zqpd\" (UniqueName: \"kubernetes.io/projected/34e6cb54-5cda-4ac4-bfe1-e59986daa349-kube-api-access-6zqpd\") pod \"dnsmasq-dns-56df8fb6b7-bdn6t\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.087026 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.213259 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zfnx8"] Oct 01 16:17:16 crc kubenswrapper[4726]: W1001 16:17:16.232684 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf1f486b_c5af_47de_9af6_4c499af830db.slice/crio-4f3e75eee6e8b3b4eb6d35a65148e7abbdf9e6dd60576f358cb465e79a6803d2 WatchSource:0}: Error finding container 4f3e75eee6e8b3b4eb6d35a65148e7abbdf9e6dd60576f358cb465e79a6803d2: Status 404 returned error can't find the container with id 4f3e75eee6e8b3b4eb6d35a65148e7abbdf9e6dd60576f358cb465e79a6803d2 Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.276575 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c6ccb6664-v52zf"] Oct 01 16:17:16 crc kubenswrapper[4726]: W1001 16:17:16.291444 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b0861ea_bb92_4592_9129_f9b3b4725a51.slice/crio-3d525627067a1f7c05e703d40e50b8f2eb250c9ba0c9e7ae96f5bf8838a4bf31 WatchSource:0}: Error finding container 3d525627067a1f7c05e703d40e50b8f2eb250c9ba0c9e7ae96f5bf8838a4bf31: Status 404 returned error can't find the container with id 3d525627067a1f7c05e703d40e50b8f2eb250c9ba0c9e7ae96f5bf8838a4bf31 Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.426539 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b5b89f8d4-d8p9g" event={"ID":"c8ee9afe-8056-4de3-bbff-524420e36290","Type":"ContainerStarted","Data":"89d40855bfc322ec70e8956bfb799f6833d982cb8089736aa2cd864b1d13c113"} Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.432652 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6ccb6664-v52zf" event={"ID":"5b0861ea-bb92-4592-9129-f9b3b4725a51","Type":"ContainerStarted","Data":"3d525627067a1f7c05e703d40e50b8f2eb250c9ba0c9e7ae96f5bf8838a4bf31"} Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.436943 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zfnx8" event={"ID":"df1f486b-c5af-47de-9af6-4c499af830db","Type":"ContainerStarted","Data":"4f3e75eee6e8b3b4eb6d35a65148e7abbdf9e6dd60576f358cb465e79a6803d2"} Oct 01 16:17:16 crc kubenswrapper[4726]: E1001 16:17:16.460762 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-gsljc" podUID="a83273b9-e1f0-47e6-8bbe-839f4c0075b3" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.566578 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bdn6t"] Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.657673 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.659029 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.660763 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-7lwk2" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.661618 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.661833 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.673760 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.834694 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/66d287be-edfc-40c8-be23-2ce44527af8a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.834777 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w867h\" (UniqueName: \"kubernetes.io/projected/66d287be-edfc-40c8-be23-2ce44527af8a-kube-api-access-w867h\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.834844 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.834916 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-scripts\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.834949 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.835082 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d287be-edfc-40c8-be23-2ce44527af8a-logs\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.835136 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-config-data\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.910867 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.913062 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.925220 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.929755 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.939204 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.939302 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-scripts\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.939330 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.939419 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d287be-edfc-40c8-be23-2ce44527af8a-logs\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.939458 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-config-data\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.939508 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/66d287be-edfc-40c8-be23-2ce44527af8a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.939600 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.946205 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-scripts\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.952479 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.952561 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d287be-edfc-40c8-be23-2ce44527af8a-logs\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.952654 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w867h\" (UniqueName: \"kubernetes.io/projected/66d287be-edfc-40c8-be23-2ce44527af8a-kube-api-access-w867h\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.952651 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/66d287be-edfc-40c8-be23-2ce44527af8a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.956656 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-config-data\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.973392 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w867h\" (UniqueName: \"kubernetes.io/projected/66d287be-edfc-40c8-be23-2ce44527af8a-kube-api-access-w867h\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:16 crc kubenswrapper[4726]: I1001 16:17:16.986652 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.041801 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.058033 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.058168 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.058333 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1265fac-c421-4a79-b148-ade1b634b6ac-logs\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.058409 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f1265fac-c421-4a79-b148-ade1b634b6ac-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.058486 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsd87\" (UniqueName: \"kubernetes.io/projected/f1265fac-c421-4a79-b148-ade1b634b6ac-kube-api-access-rsd87\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.058547 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.058580 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.159885 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1265fac-c421-4a79-b148-ade1b634b6ac-logs\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.160299 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f1265fac-c421-4a79-b148-ade1b634b6ac-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.160347 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsd87\" (UniqueName: \"kubernetes.io/projected/f1265fac-c421-4a79-b148-ade1b634b6ac-kube-api-access-rsd87\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.160373 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.160393 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.160435 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.160464 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.160685 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.160805 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1265fac-c421-4a79-b148-ade1b634b6ac-logs\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.161369 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f1265fac-c421-4a79-b148-ade1b634b6ac-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.165354 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.167176 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.174105 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.178035 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsd87\" (UniqueName: \"kubernetes.io/projected/f1265fac-c421-4a79-b148-ade1b634b6ac-kube-api-access-rsd87\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.201516 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.229623 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:17 crc kubenswrapper[4726]: W1001 16:17:17.240194 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34e6cb54_5cda_4ac4_bfe1_e59986daa349.slice/crio-bbb9783fb8558f7a4742d4e955d0800fcdeec69656503992055bbba90628b288 WatchSource:0}: Error finding container bbb9783fb8558f7a4742d4e955d0800fcdeec69656503992055bbba90628b288: Status 404 returned error can't find the container with id bbb9783fb8558f7a4742d4e955d0800fcdeec69656503992055bbba90628b288 Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.457387 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fc7f8d489-j7grp" event={"ID":"62a37356-6f11-4c75-a452-570a847a0518","Type":"ContainerStarted","Data":"d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d"} Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.459679 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b5b89f8d4-d8p9g" event={"ID":"c8ee9afe-8056-4de3-bbff-524420e36290","Type":"ContainerStarted","Data":"c2007ce7a06d7a33c05076a9c3eb97e576f90edebb480691ee958a4c79ed616d"} Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.461397 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ks2wk" event={"ID":"9c7939fb-33c3-486d-85a6-05265f29d3e8","Type":"ContainerStarted","Data":"09be41a89fb7158684759b4ae7e3644120e39d857595421392f69ecee43fd0d4"} Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.465757 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6ccb6664-v52zf" event={"ID":"5b0861ea-bb92-4592-9129-f9b3b4725a51","Type":"ContainerStarted","Data":"e870f4b47f8f55154e36b3b3cddea13fdb03c3983010f39eb6492d34869cb160"} Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.467653 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69b99cdcfc-z8t52" event={"ID":"4cff8692-89be-4a73-b714-1cc1e23a6b7d","Type":"ContainerStarted","Data":"7e72b38cef22da8024b677ace27f67c37f8e798b5cd5134a36f6376bde1188f8"} Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.470025 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7945d7f85f-2qfqm" event={"ID":"a2911957-529c-46a4-8b63-e98beb36809b","Type":"ContainerStarted","Data":"c0a9b35e55b27be633cb455b9f05dd6fdbf7f9323e7c364c5a34d99a677bf3cb"} Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.471483 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zfnx8" event={"ID":"df1f486b-c5af-47de-9af6-4c499af830db","Type":"ContainerStarted","Data":"ea492a4fbf49900bab734a0b934d107aece9b120a65c85bd355aff4831b1d322"} Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.474698 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" event={"ID":"34e6cb54-5cda-4ac4-bfe1-e59986daa349","Type":"ContainerStarted","Data":"bbb9783fb8558f7a4742d4e955d0800fcdeec69656503992055bbba90628b288"} Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.490555 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-ks2wk" podStartSLOduration=6.747041561 podStartE2EDuration="45.490529348s" podCreationTimestamp="2025-10-01 16:16:32 +0000 UTC" firstStartedPulling="2025-10-01 16:16:36.752248587 +0000 UTC m=+1049.653801164" lastFinishedPulling="2025-10-01 16:17:15.495736364 +0000 UTC m=+1088.397288951" observedRunningTime="2025-10-01 16:17:17.481188261 +0000 UTC m=+1090.382740838" watchObservedRunningTime="2025-10-01 16:17:17.490529348 +0000 UTC m=+1090.392081935" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.503920 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-zfnx8" podStartSLOduration=22.50389998 podStartE2EDuration="22.50389998s" podCreationTimestamp="2025-10-01 16:16:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:17:17.50320292 +0000 UTC m=+1090.404755497" watchObservedRunningTime="2025-10-01 16:17:17.50389998 +0000 UTC m=+1090.405452557" Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.841772 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:17:17 crc kubenswrapper[4726]: W1001 16:17:17.880802 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1265fac_c421_4a79_b148_ade1b634b6ac.slice/crio-df0f0823c5406adb9b1b180c020575a313d7fba763ac65650218568f4c7f700f WatchSource:0}: Error finding container df0f0823c5406adb9b1b180c020575a313d7fba763ac65650218568f4c7f700f: Status 404 returned error can't find the container with id df0f0823c5406adb9b1b180c020575a313d7fba763ac65650218568f4c7f700f Oct 01 16:17:17 crc kubenswrapper[4726]: I1001 16:17:17.971749 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.489715 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" event={"ID":"34e6cb54-5cda-4ac4-bfe1-e59986daa349","Type":"ContainerStarted","Data":"833925b048845f5fafb64f4ad02060e065c312c00e235d9df20ec90a80631351"} Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.496376 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395cb84c-684a-48ec-aa24-03f5a7f218ff","Type":"ContainerStarted","Data":"7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5"} Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.497652 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f1265fac-c421-4a79-b148-ade1b634b6ac","Type":"ContainerStarted","Data":"df0f0823c5406adb9b1b180c020575a313d7fba763ac65650218568f4c7f700f"} Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.500031 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6ccb6664-v52zf" event={"ID":"5b0861ea-bb92-4592-9129-f9b3b4725a51","Type":"ContainerStarted","Data":"fd9d715e9acaeb8f62267eb263a5863c5e412082eab1db4177fd27caa439f20a"} Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.515255 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69b99cdcfc-z8t52" event={"ID":"4cff8692-89be-4a73-b714-1cc1e23a6b7d","Type":"ContainerStarted","Data":"258d4058f1e0c975c4b8b6d095a9ac22bfd17548d3b7a3a21debf9ab8df075dc"} Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.515397 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-69b99cdcfc-z8t52" podUID="4cff8692-89be-4a73-b714-1cc1e23a6b7d" containerName="horizon" containerID="cri-o://258d4058f1e0c975c4b8b6d095a9ac22bfd17548d3b7a3a21debf9ab8df075dc" gracePeriod=30 Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.515373 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-69b99cdcfc-z8t52" podUID="4cff8692-89be-4a73-b714-1cc1e23a6b7d" containerName="horizon-log" containerID="cri-o://7e72b38cef22da8024b677ace27f67c37f8e798b5cd5134a36f6376bde1188f8" gracePeriod=30 Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.532587 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-c6ccb6664-v52zf" podStartSLOduration=37.532568394 podStartE2EDuration="37.532568394s" podCreationTimestamp="2025-10-01 16:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:17:18.529533647 +0000 UTC m=+1091.431086234" watchObservedRunningTime="2025-10-01 16:17:18.532568394 +0000 UTC m=+1091.434120971" Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.537432 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7945d7f85f-2qfqm" event={"ID":"a2911957-529c-46a4-8b63-e98beb36809b","Type":"ContainerStarted","Data":"ddda2e2ff2cccd5cef2be818e15e4aa658a852504f64ec8cabb4a25f0744877c"} Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.537628 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7945d7f85f-2qfqm" podUID="a2911957-529c-46a4-8b63-e98beb36809b" containerName="horizon-log" containerID="cri-o://c0a9b35e55b27be633cb455b9f05dd6fdbf7f9323e7c364c5a34d99a677bf3cb" gracePeriod=30 Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.537738 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7945d7f85f-2qfqm" podUID="a2911957-529c-46a4-8b63-e98beb36809b" containerName="horizon" containerID="cri-o://ddda2e2ff2cccd5cef2be818e15e4aa658a852504f64ec8cabb4a25f0744877c" gracePeriod=30 Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.541739 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"66d287be-edfc-40c8-be23-2ce44527af8a","Type":"ContainerStarted","Data":"21eef9763175350e7c237ebd58d1f3bf9564b6b5ed90e052533e8c26313e9f57"} Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.549326 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fc7f8d489-j7grp" event={"ID":"62a37356-6f11-4c75-a452-570a847a0518","Type":"ContainerStarted","Data":"dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9"} Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.549436 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-fc7f8d489-j7grp" podUID="62a37356-6f11-4c75-a452-570a847a0518" containerName="horizon-log" containerID="cri-o://d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d" gracePeriod=30 Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.549515 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-fc7f8d489-j7grp" podUID="62a37356-6f11-4c75-a452-570a847a0518" containerName="horizon" containerID="cri-o://dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9" gracePeriod=30 Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.552665 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b5b89f8d4-d8p9g" event={"ID":"c8ee9afe-8056-4de3-bbff-524420e36290","Type":"ContainerStarted","Data":"fe86f658aa9dd1fa4641eaea3b2cd4d5207c79815970138c88ffed0c762b1005"} Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.560936 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-69b99cdcfc-z8t52" podStartSLOduration=5.659870016 podStartE2EDuration="44.560907214s" podCreationTimestamp="2025-10-01 16:16:34 +0000 UTC" firstStartedPulling="2025-10-01 16:16:36.594697176 +0000 UTC m=+1049.496249753" lastFinishedPulling="2025-10-01 16:17:15.495734374 +0000 UTC m=+1088.397286951" observedRunningTime="2025-10-01 16:17:18.549398435 +0000 UTC m=+1091.450951022" watchObservedRunningTime="2025-10-01 16:17:18.560907214 +0000 UTC m=+1091.462459791" Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.582714 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-fc7f8d489-j7grp" podStartSLOduration=6.404128144 podStartE2EDuration="46.582689376s" podCreationTimestamp="2025-10-01 16:16:32 +0000 UTC" firstStartedPulling="2025-10-01 16:16:36.049001628 +0000 UTC m=+1048.950554195" lastFinishedPulling="2025-10-01 16:17:16.22756285 +0000 UTC m=+1089.129115427" observedRunningTime="2025-10-01 16:17:18.566694089 +0000 UTC m=+1091.468246696" watchObservedRunningTime="2025-10-01 16:17:18.582689376 +0000 UTC m=+1091.484241953" Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.608820 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7945d7f85f-2qfqm" podStartSLOduration=4.235734672 podStartE2EDuration="46.608798242s" podCreationTimestamp="2025-10-01 16:16:32 +0000 UTC" firstStartedPulling="2025-10-01 16:16:33.306148545 +0000 UTC m=+1046.207701122" lastFinishedPulling="2025-10-01 16:17:15.679212115 +0000 UTC m=+1088.580764692" observedRunningTime="2025-10-01 16:17:18.592563818 +0000 UTC m=+1091.494116395" watchObservedRunningTime="2025-10-01 16:17:18.608798242 +0000 UTC m=+1091.510350819" Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.638803 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7b5b89f8d4-d8p9g" podStartSLOduration=37.638782018 podStartE2EDuration="37.638782018s" podCreationTimestamp="2025-10-01 16:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:17:18.614597457 +0000 UTC m=+1091.516150034" watchObservedRunningTime="2025-10-01 16:17:18.638782018 +0000 UTC m=+1091.540334595" Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.655277 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:17:18 crc kubenswrapper[4726]: I1001 16:17:18.724464 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:17:19 crc kubenswrapper[4726]: I1001 16:17:19.570092 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"66d287be-edfc-40c8-be23-2ce44527af8a","Type":"ContainerStarted","Data":"900a97763a64952166995f3d2d97c0466fd11a3a256e59766c0bed6b95e2dadc"} Oct 01 16:17:19 crc kubenswrapper[4726]: I1001 16:17:19.577501 4726 generic.go:334] "Generic (PLEG): container finished" podID="34e6cb54-5cda-4ac4-bfe1-e59986daa349" containerID="833925b048845f5fafb64f4ad02060e065c312c00e235d9df20ec90a80631351" exitCode=0 Oct 01 16:17:19 crc kubenswrapper[4726]: I1001 16:17:19.577618 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" event={"ID":"34e6cb54-5cda-4ac4-bfe1-e59986daa349","Type":"ContainerDied","Data":"833925b048845f5fafb64f4ad02060e065c312c00e235d9df20ec90a80631351"} Oct 01 16:17:19 crc kubenswrapper[4726]: I1001 16:17:19.588929 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f1265fac-c421-4a79-b148-ade1b634b6ac","Type":"ContainerStarted","Data":"860aef35323df98dd89a9c5ecf2e34f6f188dad9b88a350d29ada41aa73cfd81"} Oct 01 16:17:19 crc kubenswrapper[4726]: I1001 16:17:19.589006 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f1265fac-c421-4a79-b148-ade1b634b6ac","Type":"ContainerStarted","Data":"fa4b50aa28779a7030f485be82a85c596b1e1b5ace3b868f472da024d437f4ef"} Oct 01 16:17:19 crc kubenswrapper[4726]: I1001 16:17:19.589010 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f1265fac-c421-4a79-b148-ade1b634b6ac" containerName="glance-log" containerID="cri-o://fa4b50aa28779a7030f485be82a85c596b1e1b5ace3b868f472da024d437f4ef" gracePeriod=30 Oct 01 16:17:19 crc kubenswrapper[4726]: I1001 16:17:19.589177 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f1265fac-c421-4a79-b148-ade1b634b6ac" containerName="glance-httpd" containerID="cri-o://860aef35323df98dd89a9c5ecf2e34f6f188dad9b88a350d29ada41aa73cfd81" gracePeriod=30 Oct 01 16:17:19 crc kubenswrapper[4726]: I1001 16:17:19.645785 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.645759144 podStartE2EDuration="4.645759144s" podCreationTimestamp="2025-10-01 16:17:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:17:19.636362535 +0000 UTC m=+1092.537915152" watchObservedRunningTime="2025-10-01 16:17:19.645759144 +0000 UTC m=+1092.547311721" Oct 01 16:17:20 crc kubenswrapper[4726]: I1001 16:17:20.602140 4726 generic.go:334] "Generic (PLEG): container finished" podID="f1265fac-c421-4a79-b148-ade1b634b6ac" containerID="860aef35323df98dd89a9c5ecf2e34f6f188dad9b88a350d29ada41aa73cfd81" exitCode=143 Oct 01 16:17:20 crc kubenswrapper[4726]: I1001 16:17:20.602387 4726 generic.go:334] "Generic (PLEG): container finished" podID="f1265fac-c421-4a79-b148-ade1b634b6ac" containerID="fa4b50aa28779a7030f485be82a85c596b1e1b5ace3b868f472da024d437f4ef" exitCode=143 Oct 01 16:17:20 crc kubenswrapper[4726]: I1001 16:17:20.602233 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f1265fac-c421-4a79-b148-ade1b634b6ac","Type":"ContainerDied","Data":"860aef35323df98dd89a9c5ecf2e34f6f188dad9b88a350d29ada41aa73cfd81"} Oct 01 16:17:20 crc kubenswrapper[4726]: I1001 16:17:20.602479 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f1265fac-c421-4a79-b148-ade1b634b6ac","Type":"ContainerDied","Data":"fa4b50aa28779a7030f485be82a85c596b1e1b5ace3b868f472da024d437f4ef"} Oct 01 16:17:20 crc kubenswrapper[4726]: I1001 16:17:20.607090 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"66d287be-edfc-40c8-be23-2ce44527af8a","Type":"ContainerStarted","Data":"42be7ee4e31fdf3f8dd88cde48115931ef0a716d21e2846559fedede9c6f076a"} Oct 01 16:17:22 crc kubenswrapper[4726]: I1001 16:17:22.068358 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:17:22 crc kubenswrapper[4726]: I1001 16:17:22.068849 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:17:22 crc kubenswrapper[4726]: I1001 16:17:22.321804 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:17:22 crc kubenswrapper[4726]: I1001 16:17:22.321871 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:17:22 crc kubenswrapper[4726]: I1001 16:17:22.544822 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:17:22 crc kubenswrapper[4726]: I1001 16:17:22.741346 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:17:23 crc kubenswrapper[4726]: I1001 16:17:23.635646 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" event={"ID":"34e6cb54-5cda-4ac4-bfe1-e59986daa349","Type":"ContainerStarted","Data":"c82e46e2d4b5502528f154f2d29fabd0fe75b2be4ccdfcaee912aad6362e7885"} Oct 01 16:17:23 crc kubenswrapper[4726]: I1001 16:17:23.635847 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="66d287be-edfc-40c8-be23-2ce44527af8a" containerName="glance-httpd" containerID="cri-o://42be7ee4e31fdf3f8dd88cde48115931ef0a716d21e2846559fedede9c6f076a" gracePeriod=30 Oct 01 16:17:23 crc kubenswrapper[4726]: I1001 16:17:23.636636 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="66d287be-edfc-40c8-be23-2ce44527af8a" containerName="glance-log" containerID="cri-o://900a97763a64952166995f3d2d97c0466fd11a3a256e59766c0bed6b95e2dadc" gracePeriod=30 Oct 01 16:17:23 crc kubenswrapper[4726]: I1001 16:17:23.681073 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" podStartSLOduration=8.681024596 podStartE2EDuration="8.681024596s" podCreationTimestamp="2025-10-01 16:17:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:17:23.676029643 +0000 UTC m=+1096.577582250" watchObservedRunningTime="2025-10-01 16:17:23.681024596 +0000 UTC m=+1096.582577173" Oct 01 16:17:23 crc kubenswrapper[4726]: I1001 16:17:23.701362 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.701175311 podStartE2EDuration="8.701175311s" podCreationTimestamp="2025-10-01 16:17:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:17:23.697197918 +0000 UTC m=+1096.598750675" watchObservedRunningTime="2025-10-01 16:17:23.701175311 +0000 UTC m=+1096.602727888" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.639523 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.650897 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.650902 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f1265fac-c421-4a79-b148-ade1b634b6ac","Type":"ContainerDied","Data":"df0f0823c5406adb9b1b180c020575a313d7fba763ac65650218568f4c7f700f"} Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.650980 4726 scope.go:117] "RemoveContainer" containerID="860aef35323df98dd89a9c5ecf2e34f6f188dad9b88a350d29ada41aa73cfd81" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.653433 4726 generic.go:334] "Generic (PLEG): container finished" podID="df1f486b-c5af-47de-9af6-4c499af830db" containerID="ea492a4fbf49900bab734a0b934d107aece9b120a65c85bd355aff4831b1d322" exitCode=0 Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.653511 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zfnx8" event={"ID":"df1f486b-c5af-47de-9af6-4c499af830db","Type":"ContainerDied","Data":"ea492a4fbf49900bab734a0b934d107aece9b120a65c85bd355aff4831b1d322"} Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.658002 4726 generic.go:334] "Generic (PLEG): container finished" podID="66d287be-edfc-40c8-be23-2ce44527af8a" containerID="42be7ee4e31fdf3f8dd88cde48115931ef0a716d21e2846559fedede9c6f076a" exitCode=0 Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.658058 4726 generic.go:334] "Generic (PLEG): container finished" podID="66d287be-edfc-40c8-be23-2ce44527af8a" containerID="900a97763a64952166995f3d2d97c0466fd11a3a256e59766c0bed6b95e2dadc" exitCode=143 Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.658079 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"66d287be-edfc-40c8-be23-2ce44527af8a","Type":"ContainerDied","Data":"42be7ee4e31fdf3f8dd88cde48115931ef0a716d21e2846559fedede9c6f076a"} Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.658141 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"66d287be-edfc-40c8-be23-2ce44527af8a","Type":"ContainerDied","Data":"900a97763a64952166995f3d2d97c0466fd11a3a256e59766c0bed6b95e2dadc"} Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.658409 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.735911 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-config-data\") pod \"f1265fac-c421-4a79-b148-ade1b634b6ac\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.735980 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"f1265fac-c421-4a79-b148-ade1b634b6ac\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.736006 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-scripts\") pod \"f1265fac-c421-4a79-b148-ade1b634b6ac\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.736038 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1265fac-c421-4a79-b148-ade1b634b6ac-logs\") pod \"f1265fac-c421-4a79-b148-ade1b634b6ac\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.736086 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsd87\" (UniqueName: \"kubernetes.io/projected/f1265fac-c421-4a79-b148-ade1b634b6ac-kube-api-access-rsd87\") pod \"f1265fac-c421-4a79-b148-ade1b634b6ac\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.736149 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-combined-ca-bundle\") pod \"f1265fac-c421-4a79-b148-ade1b634b6ac\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.736260 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f1265fac-c421-4a79-b148-ade1b634b6ac-httpd-run\") pod \"f1265fac-c421-4a79-b148-ade1b634b6ac\" (UID: \"f1265fac-c421-4a79-b148-ade1b634b6ac\") " Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.738499 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1265fac-c421-4a79-b148-ade1b634b6ac-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f1265fac-c421-4a79-b148-ade1b634b6ac" (UID: "f1265fac-c421-4a79-b148-ade1b634b6ac"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.738942 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1265fac-c421-4a79-b148-ade1b634b6ac-logs" (OuterVolumeSpecName: "logs") pod "f1265fac-c421-4a79-b148-ade1b634b6ac" (UID: "f1265fac-c421-4a79-b148-ade1b634b6ac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.743548 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "f1265fac-c421-4a79-b148-ade1b634b6ac" (UID: "f1265fac-c421-4a79-b148-ade1b634b6ac"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.744372 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-scripts" (OuterVolumeSpecName: "scripts") pod "f1265fac-c421-4a79-b148-ade1b634b6ac" (UID: "f1265fac-c421-4a79-b148-ade1b634b6ac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.750955 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1265fac-c421-4a79-b148-ade1b634b6ac-kube-api-access-rsd87" (OuterVolumeSpecName: "kube-api-access-rsd87") pod "f1265fac-c421-4a79-b148-ade1b634b6ac" (UID: "f1265fac-c421-4a79-b148-ade1b634b6ac"). InnerVolumeSpecName "kube-api-access-rsd87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.789974 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1265fac-c421-4a79-b148-ade1b634b6ac" (UID: "f1265fac-c421-4a79-b148-ade1b634b6ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.825186 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-config-data" (OuterVolumeSpecName: "config-data") pod "f1265fac-c421-4a79-b148-ade1b634b6ac" (UID: "f1265fac-c421-4a79-b148-ade1b634b6ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.838252 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.838325 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.838340 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.838353 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1265fac-c421-4a79-b148-ade1b634b6ac-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.838365 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsd87\" (UniqueName: \"kubernetes.io/projected/f1265fac-c421-4a79-b148-ade1b634b6ac-kube-api-access-rsd87\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.838380 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1265fac-c421-4a79-b148-ade1b634b6ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.838391 4726 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f1265fac-c421-4a79-b148-ade1b634b6ac-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.877236 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.940375 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.989543 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:17:24 crc kubenswrapper[4726]: I1001 16:17:24.998850 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.020537 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:17:25 crc kubenswrapper[4726]: E1001 16:17:25.020956 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1265fac-c421-4a79-b148-ade1b634b6ac" containerName="glance-log" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.020979 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1265fac-c421-4a79-b148-ade1b634b6ac" containerName="glance-log" Oct 01 16:17:25 crc kubenswrapper[4726]: E1001 16:17:25.021018 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1265fac-c421-4a79-b148-ade1b634b6ac" containerName="glance-httpd" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.021026 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1265fac-c421-4a79-b148-ade1b634b6ac" containerName="glance-httpd" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.021198 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1265fac-c421-4a79-b148-ade1b634b6ac" containerName="glance-httpd" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.021228 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1265fac-c421-4a79-b148-ade1b634b6ac" containerName="glance-log" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.022147 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.026560 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.026807 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.026896 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.144184 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c4283abe-f876-41c5-aa74-348a7396091f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.144276 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.144321 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.144347 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltk56\" (UniqueName: \"kubernetes.io/projected/c4283abe-f876-41c5-aa74-348a7396091f-kube-api-access-ltk56\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.144379 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.144399 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.144419 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4283abe-f876-41c5-aa74-348a7396091f-logs\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.144439 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.151554 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.246307 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.246400 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.246437 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltk56\" (UniqueName: \"kubernetes.io/projected/c4283abe-f876-41c5-aa74-348a7396091f-kube-api-access-ltk56\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.246491 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.246522 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.246549 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4283abe-f876-41c5-aa74-348a7396091f-logs\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.246580 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.246641 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c4283abe-f876-41c5-aa74-348a7396091f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.247203 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c4283abe-f876-41c5-aa74-348a7396091f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.247439 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.248975 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4283abe-f876-41c5-aa74-348a7396091f-logs\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.261761 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.261840 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.265072 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.265436 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.267844 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltk56\" (UniqueName: \"kubernetes.io/projected/c4283abe-f876-41c5-aa74-348a7396091f-kube-api-access-ltk56\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.313277 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.377836 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:25 crc kubenswrapper[4726]: I1001 16:17:25.852154 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1265fac-c421-4a79-b148-ade1b634b6ac" path="/var/lib/kubelet/pods/f1265fac-c421-4a79-b148-ade1b634b6ac/volumes" Oct 01 16:17:26 crc kubenswrapper[4726]: I1001 16:17:26.654202 4726 scope.go:117] "RemoveContainer" containerID="fa4b50aa28779a7030f485be82a85c596b1e1b5ace3b868f472da024d437f4ef" Oct 01 16:17:26 crc kubenswrapper[4726]: I1001 16:17:26.690082 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zfnx8" event={"ID":"df1f486b-c5af-47de-9af6-4c499af830db","Type":"ContainerDied","Data":"4f3e75eee6e8b3b4eb6d35a65148e7abbdf9e6dd60576f358cb465e79a6803d2"} Oct 01 16:17:26 crc kubenswrapper[4726]: I1001 16:17:26.690203 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f3e75eee6e8b3b4eb6d35a65148e7abbdf9e6dd60576f358cb465e79a6803d2" Oct 01 16:17:26 crc kubenswrapper[4726]: I1001 16:17:26.692169 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"66d287be-edfc-40c8-be23-2ce44527af8a","Type":"ContainerDied","Data":"21eef9763175350e7c237ebd58d1f3bf9564b6b5ed90e052533e8c26313e9f57"} Oct 01 16:17:26 crc kubenswrapper[4726]: I1001 16:17:26.692245 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21eef9763175350e7c237ebd58d1f3bf9564b6b5ed90e052533e8c26313e9f57" Oct 01 16:17:26 crc kubenswrapper[4726]: I1001 16:17:26.693407 4726 generic.go:334] "Generic (PLEG): container finished" podID="9c7939fb-33c3-486d-85a6-05265f29d3e8" containerID="09be41a89fb7158684759b4ae7e3644120e39d857595421392f69ecee43fd0d4" exitCode=0 Oct 01 16:17:26 crc kubenswrapper[4726]: I1001 16:17:26.693445 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ks2wk" event={"ID":"9c7939fb-33c3-486d-85a6-05265f29d3e8","Type":"ContainerDied","Data":"09be41a89fb7158684759b4ae7e3644120e39d857595421392f69ecee43fd0d4"} Oct 01 16:17:26 crc kubenswrapper[4726]: I1001 16:17:26.956441 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:17:26 crc kubenswrapper[4726]: I1001 16:17:26.963788 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.052372 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:17:27 crc kubenswrapper[4726]: W1001 16:17:27.064426 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4283abe_f876_41c5_aa74_348a7396091f.slice/crio-b446ee5c74757b35755688e1eee3b7122b75ca3a1cc4c1acfff8ecb0ac8d22c2 WatchSource:0}: Error finding container b446ee5c74757b35755688e1eee3b7122b75ca3a1cc4c1acfff8ecb0ac8d22c2: Status 404 returned error can't find the container with id b446ee5c74757b35755688e1eee3b7122b75ca3a1cc4c1acfff8ecb0ac8d22c2 Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.104731 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d287be-edfc-40c8-be23-2ce44527af8a-logs\") pod \"66d287be-edfc-40c8-be23-2ce44527af8a\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.104811 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-fernet-keys\") pod \"df1f486b-c5af-47de-9af6-4c499af830db\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.104866 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/66d287be-edfc-40c8-be23-2ce44527af8a-httpd-run\") pod \"66d287be-edfc-40c8-be23-2ce44527af8a\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.104910 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-credential-keys\") pod \"df1f486b-c5af-47de-9af6-4c499af830db\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.104949 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-scripts\") pod \"66d287be-edfc-40c8-be23-2ce44527af8a\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.105001 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w867h\" (UniqueName: \"kubernetes.io/projected/66d287be-edfc-40c8-be23-2ce44527af8a-kube-api-access-w867h\") pod \"66d287be-edfc-40c8-be23-2ce44527af8a\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.105065 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89pjw\" (UniqueName: \"kubernetes.io/projected/df1f486b-c5af-47de-9af6-4c499af830db-kube-api-access-89pjw\") pod \"df1f486b-c5af-47de-9af6-4c499af830db\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.105113 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"66d287be-edfc-40c8-be23-2ce44527af8a\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.105133 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-config-data\") pod \"df1f486b-c5af-47de-9af6-4c499af830db\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.105150 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-config-data\") pod \"66d287be-edfc-40c8-be23-2ce44527af8a\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.105166 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-combined-ca-bundle\") pod \"66d287be-edfc-40c8-be23-2ce44527af8a\" (UID: \"66d287be-edfc-40c8-be23-2ce44527af8a\") " Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.105224 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-scripts\") pod \"df1f486b-c5af-47de-9af6-4c499af830db\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.105222 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66d287be-edfc-40c8-be23-2ce44527af8a-logs" (OuterVolumeSpecName: "logs") pod "66d287be-edfc-40c8-be23-2ce44527af8a" (UID: "66d287be-edfc-40c8-be23-2ce44527af8a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.105242 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-combined-ca-bundle\") pod \"df1f486b-c5af-47de-9af6-4c499af830db\" (UID: \"df1f486b-c5af-47de-9af6-4c499af830db\") " Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.105580 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d287be-edfc-40c8-be23-2ce44527af8a-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.107642 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66d287be-edfc-40c8-be23-2ce44527af8a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "66d287be-edfc-40c8-be23-2ce44527af8a" (UID: "66d287be-edfc-40c8-be23-2ce44527af8a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.114201 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66d287be-edfc-40c8-be23-2ce44527af8a-kube-api-access-w867h" (OuterVolumeSpecName: "kube-api-access-w867h") pod "66d287be-edfc-40c8-be23-2ce44527af8a" (UID: "66d287be-edfc-40c8-be23-2ce44527af8a"). InnerVolumeSpecName "kube-api-access-w867h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.115196 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-scripts" (OuterVolumeSpecName: "scripts") pod "66d287be-edfc-40c8-be23-2ce44527af8a" (UID: "66d287be-edfc-40c8-be23-2ce44527af8a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.115239 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "df1f486b-c5af-47de-9af6-4c499af830db" (UID: "df1f486b-c5af-47de-9af6-4c499af830db"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.115298 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df1f486b-c5af-47de-9af6-4c499af830db-kube-api-access-89pjw" (OuterVolumeSpecName: "kube-api-access-89pjw") pod "df1f486b-c5af-47de-9af6-4c499af830db" (UID: "df1f486b-c5af-47de-9af6-4c499af830db"). InnerVolumeSpecName "kube-api-access-89pjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.115309 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "df1f486b-c5af-47de-9af6-4c499af830db" (UID: "df1f486b-c5af-47de-9af6-4c499af830db"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.123115 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-scripts" (OuterVolumeSpecName: "scripts") pod "df1f486b-c5af-47de-9af6-4c499af830db" (UID: "df1f486b-c5af-47de-9af6-4c499af830db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.131191 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "66d287be-edfc-40c8-be23-2ce44527af8a" (UID: "66d287be-edfc-40c8-be23-2ce44527af8a"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.133219 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df1f486b-c5af-47de-9af6-4c499af830db" (UID: "df1f486b-c5af-47de-9af6-4c499af830db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.163340 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66d287be-edfc-40c8-be23-2ce44527af8a" (UID: "66d287be-edfc-40c8-be23-2ce44527af8a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.170424 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-config-data" (OuterVolumeSpecName: "config-data") pod "df1f486b-c5af-47de-9af6-4c499af830db" (UID: "df1f486b-c5af-47de-9af6-4c499af830db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.178187 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-config-data" (OuterVolumeSpecName: "config-data") pod "66d287be-edfc-40c8-be23-2ce44527af8a" (UID: "66d287be-edfc-40c8-be23-2ce44527af8a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.207392 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.207448 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.207464 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.207477 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.207492 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.207504 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.207515 4726 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.207526 4726 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/66d287be-edfc-40c8-be23-2ce44527af8a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.207536 4726 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/df1f486b-c5af-47de-9af6-4c499af830db-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.207548 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66d287be-edfc-40c8-be23-2ce44527af8a-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.207559 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w867h\" (UniqueName: \"kubernetes.io/projected/66d287be-edfc-40c8-be23-2ce44527af8a-kube-api-access-w867h\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.207571 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89pjw\" (UniqueName: \"kubernetes.io/projected/df1f486b-c5af-47de-9af6-4c499af830db-kube-api-access-89pjw\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.235416 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.309025 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.716939 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395cb84c-684a-48ec-aa24-03f5a7f218ff","Type":"ContainerStarted","Data":"698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855"} Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.719772 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c4283abe-f876-41c5-aa74-348a7396091f","Type":"ContainerStarted","Data":"c2aa66ea86e3e5a10e1d812a1c8294b844da3d725de401714373e9be0e2200a1"} Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.719795 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c4283abe-f876-41c5-aa74-348a7396091f","Type":"ContainerStarted","Data":"b446ee5c74757b35755688e1eee3b7122b75ca3a1cc4c1acfff8ecb0ac8d22c2"} Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.723228 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.723481 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zfnx8" Oct 01 16:17:27 crc kubenswrapper[4726]: I1001 16:17:27.963567 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.025567 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.040042 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:17:28 crc kubenswrapper[4726]: E1001 16:17:28.043531 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d287be-edfc-40c8-be23-2ce44527af8a" containerName="glance-httpd" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.043554 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d287be-edfc-40c8-be23-2ce44527af8a" containerName="glance-httpd" Oct 01 16:17:28 crc kubenswrapper[4726]: E1001 16:17:28.043580 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df1f486b-c5af-47de-9af6-4c499af830db" containerName="keystone-bootstrap" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.043587 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="df1f486b-c5af-47de-9af6-4c499af830db" containerName="keystone-bootstrap" Oct 01 16:17:28 crc kubenswrapper[4726]: E1001 16:17:28.043610 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d287be-edfc-40c8-be23-2ce44527af8a" containerName="glance-log" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.043616 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d287be-edfc-40c8-be23-2ce44527af8a" containerName="glance-log" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.043803 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="df1f486b-c5af-47de-9af6-4c499af830db" containerName="keystone-bootstrap" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.043818 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d287be-edfc-40c8-be23-2ce44527af8a" containerName="glance-log" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.043828 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d287be-edfc-40c8-be23-2ce44527af8a" containerName="glance-httpd" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.044720 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.047627 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.047983 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.055653 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.126024 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-75bf4bdcc5-mbcpq"] Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.131239 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.133145 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4d4b770c-28b0-47c1-ba00-584e3faa804c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.133173 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d4b770c-28b0-47c1-ba00-584e3faa804c-logs\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.133236 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.133251 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.133277 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2sm7\" (UniqueName: \"kubernetes.io/projected/4d4b770c-28b0-47c1-ba00-584e3faa804c-kube-api-access-s2sm7\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.133292 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-config-data\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.133327 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-scripts\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.133341 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.143589 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.144012 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.144258 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.144558 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.144787 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5q5v6" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.144978 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.150042 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-75bf4bdcc5-mbcpq"] Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235076 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-fernet-keys\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235121 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-config-data\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235144 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-public-tls-certs\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235189 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4d4b770c-28b0-47c1-ba00-584e3faa804c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235208 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d4b770c-28b0-47c1-ba00-584e3faa804c-logs\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235234 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-internal-tls-certs\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235260 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-scripts\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235281 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-combined-ca-bundle\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235304 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-credential-keys\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235327 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m54f8\" (UniqueName: \"kubernetes.io/projected/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-kube-api-access-m54f8\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235351 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235367 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235391 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2sm7\" (UniqueName: \"kubernetes.io/projected/4d4b770c-28b0-47c1-ba00-584e3faa804c-kube-api-access-s2sm7\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235407 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-config-data\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235444 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-scripts\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.235457 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.241943 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4d4b770c-28b0-47c1-ba00-584e3faa804c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.242213 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d4b770c-28b0-47c1-ba00-584e3faa804c-logs\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.242616 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.248924 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.249525 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-config-data\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.250625 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.258767 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-scripts\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.277882 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2sm7\" (UniqueName: \"kubernetes.io/projected/4d4b770c-28b0-47c1-ba00-584e3faa804c-kube-api-access-s2sm7\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.283512 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ks2wk" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.300524 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.346757 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-scripts\") pod \"9c7939fb-33c3-486d-85a6-05265f29d3e8\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.346800 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c7939fb-33c3-486d-85a6-05265f29d3e8-logs\") pod \"9c7939fb-33c3-486d-85a6-05265f29d3e8\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.347000 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrq5k\" (UniqueName: \"kubernetes.io/projected/9c7939fb-33c3-486d-85a6-05265f29d3e8-kube-api-access-rrq5k\") pod \"9c7939fb-33c3-486d-85a6-05265f29d3e8\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.347028 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-combined-ca-bundle\") pod \"9c7939fb-33c3-486d-85a6-05265f29d3e8\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.347124 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-config-data\") pod \"9c7939fb-33c3-486d-85a6-05265f29d3e8\" (UID: \"9c7939fb-33c3-486d-85a6-05265f29d3e8\") " Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.347345 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m54f8\" (UniqueName: \"kubernetes.io/projected/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-kube-api-access-m54f8\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.347436 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-fernet-keys\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.347469 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-config-data\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.347490 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-public-tls-certs\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.347547 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-internal-tls-certs\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.347576 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-scripts\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.347605 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-combined-ca-bundle\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.347631 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-credential-keys\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.352316 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c7939fb-33c3-486d-85a6-05265f29d3e8-logs" (OuterVolumeSpecName: "logs") pod "9c7939fb-33c3-486d-85a6-05265f29d3e8" (UID: "9c7939fb-33c3-486d-85a6-05265f29d3e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.355306 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c7939fb-33c3-486d-85a6-05265f29d3e8-kube-api-access-rrq5k" (OuterVolumeSpecName: "kube-api-access-rrq5k") pod "9c7939fb-33c3-486d-85a6-05265f29d3e8" (UID: "9c7939fb-33c3-486d-85a6-05265f29d3e8"). InnerVolumeSpecName "kube-api-access-rrq5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.359208 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-combined-ca-bundle\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.361390 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-config-data\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.361900 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.368771 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-internal-tls-certs\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.369498 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-scripts" (OuterVolumeSpecName: "scripts") pod "9c7939fb-33c3-486d-85a6-05265f29d3e8" (UID: "9c7939fb-33c3-486d-85a6-05265f29d3e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.370245 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-credential-keys\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.370265 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-public-tls-certs\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.370378 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-scripts\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.371703 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-fernet-keys\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.382987 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m54f8\" (UniqueName: \"kubernetes.io/projected/033ea7d7-f7e9-49dd-8c50-5402d4a1aa89-kube-api-access-m54f8\") pod \"keystone-75bf4bdcc5-mbcpq\" (UID: \"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89\") " pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.386731 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-config-data" (OuterVolumeSpecName: "config-data") pod "9c7939fb-33c3-486d-85a6-05265f29d3e8" (UID: "9c7939fb-33c3-486d-85a6-05265f29d3e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.390548 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c7939fb-33c3-486d-85a6-05265f29d3e8" (UID: "9c7939fb-33c3-486d-85a6-05265f29d3e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.448791 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.449008 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.449022 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c7939fb-33c3-486d-85a6-05265f29d3e8-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.449032 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrq5k\" (UniqueName: \"kubernetes.io/projected/9c7939fb-33c3-486d-85a6-05265f29d3e8-kube-api-access-rrq5k\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.449040 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c7939fb-33c3-486d-85a6-05265f29d3e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.578481 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.783905 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ks2wk" event={"ID":"9c7939fb-33c3-486d-85a6-05265f29d3e8","Type":"ContainerDied","Data":"484bfc6ee63f2859d77f581ea837c8101abe4689fb5d35d30836f046c711ebd4"} Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.784139 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="484bfc6ee63f2859d77f581ea837c8101abe4689fb5d35d30836f046c711ebd4" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.784216 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ks2wk" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.818462 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c4283abe-f876-41c5-aa74-348a7396091f","Type":"ContainerStarted","Data":"cfbcc613ec702e703d65c0eb38d0837c1cede2e51f043c403845485988b616ca"} Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.899612 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.8995957489999995 podStartE2EDuration="4.899595749s" podCreationTimestamp="2025-10-01 16:17:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:17:28.883383946 +0000 UTC m=+1101.784936543" watchObservedRunningTime="2025-10-01 16:17:28.899595749 +0000 UTC m=+1101.801148326" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.906225 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-68f479c67b-gj9j5"] Oct 01 16:17:28 crc kubenswrapper[4726]: E1001 16:17:28.907808 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c7939fb-33c3-486d-85a6-05265f29d3e8" containerName="placement-db-sync" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.907832 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c7939fb-33c3-486d-85a6-05265f29d3e8" containerName="placement-db-sync" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.908017 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c7939fb-33c3-486d-85a6-05265f29d3e8" containerName="placement-db-sync" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.908919 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.914504 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.914651 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-fvv6p" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.914746 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.914843 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.914940 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.937713 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-68f479c67b-gj9j5"] Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.963101 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-combined-ca-bundle\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.963165 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg9cz\" (UniqueName: \"kubernetes.io/projected/f3ada4b2-6da4-4139-ad97-0d45034918aa-kube-api-access-qg9cz\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.963208 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-config-data\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.963230 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-internal-tls-certs\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.963270 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3ada4b2-6da4-4139-ad97-0d45034918aa-logs\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.963293 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-scripts\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:28 crc kubenswrapper[4726]: I1001 16:17:28.963323 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-public-tls-certs\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.064518 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-public-tls-certs\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.065487 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-combined-ca-bundle\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.065607 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg9cz\" (UniqueName: \"kubernetes.io/projected/f3ada4b2-6da4-4139-ad97-0d45034918aa-kube-api-access-qg9cz\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.065700 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-config-data\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.065751 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-internal-tls-certs\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.065843 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3ada4b2-6da4-4139-ad97-0d45034918aa-logs\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.065889 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-scripts\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.066543 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3ada4b2-6da4-4139-ad97-0d45034918aa-logs\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.072791 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-combined-ca-bundle\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.072935 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-config-data\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.073176 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-internal-tls-certs\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.076469 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-scripts\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.089529 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ada4b2-6da4-4139-ad97-0d45034918aa-public-tls-certs\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.093507 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg9cz\" (UniqueName: \"kubernetes.io/projected/f3ada4b2-6da4-4139-ad97-0d45034918aa-kube-api-access-qg9cz\") pod \"placement-68f479c67b-gj9j5\" (UID: \"f3ada4b2-6da4-4139-ad97-0d45034918aa\") " pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.131524 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:17:29 crc kubenswrapper[4726]: W1001 16:17:29.135291 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod033ea7d7_f7e9_49dd_8c50_5402d4a1aa89.slice/crio-5019a7f898ff20e972a82e80877c865de1105351eafc7393ee075f5ce4ea44a4 WatchSource:0}: Error finding container 5019a7f898ff20e972a82e80877c865de1105351eafc7393ee075f5ce4ea44a4: Status 404 returned error can't find the container with id 5019a7f898ff20e972a82e80877c865de1105351eafc7393ee075f5ce4ea44a4 Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.137393 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-75bf4bdcc5-mbcpq"] Oct 01 16:17:29 crc kubenswrapper[4726]: W1001 16:17:29.142288 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d4b770c_28b0_47c1_ba00_584e3faa804c.slice/crio-5f2976fabdad8e6ede7b523c8a431f4ea34a3f85cf0e39a7e293a13c40bcb790 WatchSource:0}: Error finding container 5f2976fabdad8e6ede7b523c8a431f4ea34a3f85cf0e39a7e293a13c40bcb790: Status 404 returned error can't find the container with id 5f2976fabdad8e6ede7b523c8a431f4ea34a3f85cf0e39a7e293a13c40bcb790 Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.253625 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.718521 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-68f479c67b-gj9j5"] Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.819012 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66d287be-edfc-40c8-be23-2ce44527af8a" path="/var/lib/kubelet/pods/66d287be-edfc-40c8-be23-2ce44527af8a/volumes" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.829886 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bhxd2" event={"ID":"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6","Type":"ContainerStarted","Data":"0fd712d7373def2d15480febabcafaa3ba5119977c938919054b21fd5ef60748"} Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.838098 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-75bf4bdcc5-mbcpq" event={"ID":"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89","Type":"ContainerStarted","Data":"74c80563c52963bcfd6d8b79bde3b047f0fb28d99ebd22e497419a249f73789a"} Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.838203 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-75bf4bdcc5-mbcpq" event={"ID":"033ea7d7-f7e9-49dd-8c50-5402d4a1aa89","Type":"ContainerStarted","Data":"5019a7f898ff20e972a82e80877c865de1105351eafc7393ee075f5ce4ea44a4"} Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.838284 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.840067 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68f479c67b-gj9j5" event={"ID":"f3ada4b2-6da4-4139-ad97-0d45034918aa","Type":"ContainerStarted","Data":"b11d1fd5d8bbcd51510ad442f4fec045812338f5235d3e06d5e48b7edd0cd5ce"} Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.845302 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-bhxd2" podStartSLOduration=5.195675774 podStartE2EDuration="57.845283234s" podCreationTimestamp="2025-10-01 16:16:32 +0000 UTC" firstStartedPulling="2025-10-01 16:16:36.064693736 +0000 UTC m=+1048.966246313" lastFinishedPulling="2025-10-01 16:17:28.714301196 +0000 UTC m=+1101.615853773" observedRunningTime="2025-10-01 16:17:29.84514816 +0000 UTC m=+1102.746700747" watchObservedRunningTime="2025-10-01 16:17:29.845283234 +0000 UTC m=+1102.746835811" Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.855454 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4d4b770c-28b0-47c1-ba00-584e3faa804c","Type":"ContainerStarted","Data":"5f2976fabdad8e6ede7b523c8a431f4ea34a3f85cf0e39a7e293a13c40bcb790"} Oct 01 16:17:29 crc kubenswrapper[4726]: I1001 16:17:29.905128 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-75bf4bdcc5-mbcpq" podStartSLOduration=1.900397008 podStartE2EDuration="1.900397008s" podCreationTimestamp="2025-10-01 16:17:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:17:29.889727873 +0000 UTC m=+1102.791280470" watchObservedRunningTime="2025-10-01 16:17:29.900397008 +0000 UTC m=+1102.801949585" Oct 01 16:17:30 crc kubenswrapper[4726]: I1001 16:17:30.879178 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68f479c67b-gj9j5" event={"ID":"f3ada4b2-6da4-4139-ad97-0d45034918aa","Type":"ContainerStarted","Data":"9fb693f8ced462b5f79439e11fe861ccf53d0a875cef50b22f8eab6049a362a0"} Oct 01 16:17:30 crc kubenswrapper[4726]: I1001 16:17:30.879751 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:30 crc kubenswrapper[4726]: I1001 16:17:30.879770 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68f479c67b-gj9j5" event={"ID":"f3ada4b2-6da4-4139-ad97-0d45034918aa","Type":"ContainerStarted","Data":"1070022b32bcab82b1118312f3a7122d710692acbe66c80fd564a3105e035db8"} Oct 01 16:17:30 crc kubenswrapper[4726]: I1001 16:17:30.886477 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4d4b770c-28b0-47c1-ba00-584e3faa804c","Type":"ContainerStarted","Data":"bd33518a068228fb54a715e346710f1d194a7173acb87056b88a9e850b11315a"} Oct 01 16:17:30 crc kubenswrapper[4726]: I1001 16:17:30.905942 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-68f479c67b-gj9j5" podStartSLOduration=2.9059156120000003 podStartE2EDuration="2.905915612s" podCreationTimestamp="2025-10-01 16:17:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:17:30.903824522 +0000 UTC m=+1103.805377099" watchObservedRunningTime="2025-10-01 16:17:30.905915612 +0000 UTC m=+1103.807468189" Oct 01 16:17:31 crc kubenswrapper[4726]: I1001 16:17:31.090057 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:17:31 crc kubenswrapper[4726]: I1001 16:17:31.194133 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-f594w"] Oct 01 16:17:31 crc kubenswrapper[4726]: I1001 16:17:31.194424 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf78879c9-f594w" podUID="b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" containerName="dnsmasq-dns" containerID="cri-o://aee4c133a88a3f9da53ef17d7867178dda83f566fe8fa055284ea1039ca52a6f" gracePeriod=10 Oct 01 16:17:31 crc kubenswrapper[4726]: I1001 16:17:31.899894 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gsljc" event={"ID":"a83273b9-e1f0-47e6-8bbe-839f4c0075b3","Type":"ContainerStarted","Data":"16d3ded60afe96d427098a307e71f20d4f35f12993be114dad1f6c2582f663a1"} Oct 01 16:17:31 crc kubenswrapper[4726]: I1001 16:17:31.902614 4726 generic.go:334] "Generic (PLEG): container finished" podID="b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" containerID="aee4c133a88a3f9da53ef17d7867178dda83f566fe8fa055284ea1039ca52a6f" exitCode=0 Oct 01 16:17:31 crc kubenswrapper[4726]: I1001 16:17:31.902670 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-f594w" event={"ID":"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed","Type":"ContainerDied","Data":"aee4c133a88a3f9da53ef17d7867178dda83f566fe8fa055284ea1039ca52a6f"} Oct 01 16:17:31 crc kubenswrapper[4726]: I1001 16:17:31.907522 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4d4b770c-28b0-47c1-ba00-584e3faa804c","Type":"ContainerStarted","Data":"a26e79b359150c8cdd5d683f8211f3209fc5a21fbb8c7b2213808f40d87c20b3"} Oct 01 16:17:31 crc kubenswrapper[4726]: I1001 16:17:31.907563 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:17:31 crc kubenswrapper[4726]: I1001 16:17:31.949737 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.949717079 podStartE2EDuration="4.949717079s" podCreationTimestamp="2025-10-01 16:17:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:17:31.939806427 +0000 UTC m=+1104.841359024" watchObservedRunningTime="2025-10-01 16:17:31.949717079 +0000 UTC m=+1104.851269666" Oct 01 16:17:32 crc kubenswrapper[4726]: I1001 16:17:32.079347 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c6ccb6664-v52zf" podUID="5b0861ea-bb92-4592-9129-f9b3b4725a51" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 01 16:17:32 crc kubenswrapper[4726]: I1001 16:17:32.324700 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b5b89f8d4-d8p9g" podUID="c8ee9afe-8056-4de3-bbff-524420e36290" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Oct 01 16:17:32 crc kubenswrapper[4726]: I1001 16:17:32.941038 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-gsljc" podStartSLOduration=6.780015312 podStartE2EDuration="1m0.941013217s" podCreationTimestamp="2025-10-01 16:16:32 +0000 UTC" firstStartedPulling="2025-10-01 16:16:36.060461215 +0000 UTC m=+1048.962013792" lastFinishedPulling="2025-10-01 16:17:30.22145912 +0000 UTC m=+1103.123011697" observedRunningTime="2025-10-01 16:17:32.931875346 +0000 UTC m=+1105.833427933" watchObservedRunningTime="2025-10-01 16:17:32.941013217 +0000 UTC m=+1105.842565814" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.355245 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.473236 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-config\") pod \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.473328 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-dns-swift-storage-0\") pod \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.473395 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-dns-svc\") pod \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.473413 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-ovsdbserver-sb\") pod \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.473507 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrqhr\" (UniqueName: \"kubernetes.io/projected/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-kube-api-access-nrqhr\") pod \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.473566 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-ovsdbserver-nb\") pod \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\" (UID: \"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed\") " Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.499307 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-kube-api-access-nrqhr" (OuterVolumeSpecName: "kube-api-access-nrqhr") pod "b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" (UID: "b5c407e4-2f8c-47dd-85b4-b29c12ea3aed"). InnerVolumeSpecName "kube-api-access-nrqhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.532493 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-config" (OuterVolumeSpecName: "config") pod "b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" (UID: "b5c407e4-2f8c-47dd-85b4-b29c12ea3aed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.553663 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" (UID: "b5c407e4-2f8c-47dd-85b4-b29c12ea3aed"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.554249 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" (UID: "b5c407e4-2f8c-47dd-85b4-b29c12ea3aed"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.556306 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" (UID: "b5c407e4-2f8c-47dd-85b4-b29c12ea3aed"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.557662 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" (UID: "b5c407e4-2f8c-47dd-85b4-b29c12ea3aed"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.576540 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrqhr\" (UniqueName: \"kubernetes.io/projected/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-kube-api-access-nrqhr\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.576637 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.576654 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.576665 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.576673 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.576702 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.944260 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-f594w" event={"ID":"b5c407e4-2f8c-47dd-85b4-b29c12ea3aed","Type":"ContainerDied","Data":"9c7797c6d4e7f2728b0d1cae24942ed994e8ae747c67184276ca993b91c9b787"} Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.944477 4726 scope.go:117] "RemoveContainer" containerID="aee4c133a88a3f9da53ef17d7867178dda83f566fe8fa055284ea1039ca52a6f" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.944507 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-f594w" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.976063 4726 scope.go:117] "RemoveContainer" containerID="7afdceed352942813bfe7a35b00f27efbf39f21902a199081e5e2deaabce7ee4" Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.979228 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-f594w"] Oct 01 16:17:33 crc kubenswrapper[4726]: I1001 16:17:33.994040 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-f594w"] Oct 01 16:17:35 crc kubenswrapper[4726]: I1001 16:17:35.378567 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:35 crc kubenswrapper[4726]: I1001 16:17:35.378920 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:35 crc kubenswrapper[4726]: I1001 16:17:35.416131 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:35 crc kubenswrapper[4726]: I1001 16:17:35.431486 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:35 crc kubenswrapper[4726]: I1001 16:17:35.818630 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" path="/var/lib/kubelet/pods/b5c407e4-2f8c-47dd-85b4-b29c12ea3aed/volumes" Oct 01 16:17:35 crc kubenswrapper[4726]: I1001 16:17:35.965990 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:35 crc kubenswrapper[4726]: I1001 16:17:35.966039 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:38 crc kubenswrapper[4726]: I1001 16:17:38.245670 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:38 crc kubenswrapper[4726]: I1001 16:17:38.246154 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 16:17:38 crc kubenswrapper[4726]: I1001 16:17:38.298802 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 16:17:38 crc kubenswrapper[4726]: I1001 16:17:38.365573 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 16:17:38 crc kubenswrapper[4726]: I1001 16:17:38.365935 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 16:17:38 crc kubenswrapper[4726]: I1001 16:17:38.410414 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 16:17:38 crc kubenswrapper[4726]: I1001 16:17:38.427548 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 16:17:38 crc kubenswrapper[4726]: I1001 16:17:38.988780 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 16:17:38 crc kubenswrapper[4726]: I1001 16:17:38.988819 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 16:17:40 crc kubenswrapper[4726]: I1001 16:17:40.924911 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 16:17:40 crc kubenswrapper[4726]: I1001 16:17:40.925993 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 16:17:42 crc kubenswrapper[4726]: I1001 16:17:42.069098 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c6ccb6664-v52zf" podUID="5b0861ea-bb92-4592-9129-f9b3b4725a51" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 01 16:17:44 crc kubenswrapper[4726]: I1001 16:17:44.196514 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:17:44 crc kubenswrapper[4726]: E1001 16:17:44.684501 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Oct 01 16:17:44 crc kubenswrapper[4726]: E1001 16:17:44.684716 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kjd46,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(395cb84c-684a-48ec-aa24-03f5a7f218ff): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 16:17:44 crc kubenswrapper[4726]: E1001 16:17:44.686495 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="395cb84c-684a-48ec-aa24-03f5a7f218ff" Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.045380 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="395cb84c-684a-48ec-aa24-03f5a7f218ff" containerName="ceilometer-notification-agent" containerID="cri-o://7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5" gracePeriod=30 Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.045453 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="395cb84c-684a-48ec-aa24-03f5a7f218ff" containerName="sg-core" containerID="cri-o://698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855" gracePeriod=30 Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.819788 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.900646 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7b5b89f8d4-d8p9g" Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.945149 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-config-data\") pod \"395cb84c-684a-48ec-aa24-03f5a7f218ff\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.945353 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-sg-core-conf-yaml\") pod \"395cb84c-684a-48ec-aa24-03f5a7f218ff\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.945400 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-combined-ca-bundle\") pod \"395cb84c-684a-48ec-aa24-03f5a7f218ff\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.945479 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjd46\" (UniqueName: \"kubernetes.io/projected/395cb84c-684a-48ec-aa24-03f5a7f218ff-kube-api-access-kjd46\") pod \"395cb84c-684a-48ec-aa24-03f5a7f218ff\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.945559 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-scripts\") pod \"395cb84c-684a-48ec-aa24-03f5a7f218ff\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.945600 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395cb84c-684a-48ec-aa24-03f5a7f218ff-run-httpd\") pod \"395cb84c-684a-48ec-aa24-03f5a7f218ff\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.945639 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395cb84c-684a-48ec-aa24-03f5a7f218ff-log-httpd\") pod \"395cb84c-684a-48ec-aa24-03f5a7f218ff\" (UID: \"395cb84c-684a-48ec-aa24-03f5a7f218ff\") " Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.946126 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/395cb84c-684a-48ec-aa24-03f5a7f218ff-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "395cb84c-684a-48ec-aa24-03f5a7f218ff" (UID: "395cb84c-684a-48ec-aa24-03f5a7f218ff"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.946337 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395cb84c-684a-48ec-aa24-03f5a7f218ff-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.946961 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/395cb84c-684a-48ec-aa24-03f5a7f218ff-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "395cb84c-684a-48ec-aa24-03f5a7f218ff" (UID: "395cb84c-684a-48ec-aa24-03f5a7f218ff"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.956440 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/395cb84c-684a-48ec-aa24-03f5a7f218ff-kube-api-access-kjd46" (OuterVolumeSpecName: "kube-api-access-kjd46") pod "395cb84c-684a-48ec-aa24-03f5a7f218ff" (UID: "395cb84c-684a-48ec-aa24-03f5a7f218ff"). InnerVolumeSpecName "kube-api-access-kjd46". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.968029 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c6ccb6664-v52zf"] Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.968439 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c6ccb6664-v52zf" podUID="5b0861ea-bb92-4592-9129-f9b3b4725a51" containerName="horizon" containerID="cri-o://fd9d715e9acaeb8f62267eb263a5863c5e412082eab1db4177fd27caa439f20a" gracePeriod=30 Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.970749 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c6ccb6664-v52zf" podUID="5b0861ea-bb92-4592-9129-f9b3b4725a51" containerName="horizon-log" containerID="cri-o://e870f4b47f8f55154e36b3b3cddea13fdb03c3983010f39eb6492d34869cb160" gracePeriod=30 Oct 01 16:17:45 crc kubenswrapper[4726]: I1001 16:17:45.974015 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-scripts" (OuterVolumeSpecName: "scripts") pod "395cb84c-684a-48ec-aa24-03f5a7f218ff" (UID: "395cb84c-684a-48ec-aa24-03f5a7f218ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.001256 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-config-data" (OuterVolumeSpecName: "config-data") pod "395cb84c-684a-48ec-aa24-03f5a7f218ff" (UID: "395cb84c-684a-48ec-aa24-03f5a7f218ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.008262 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "395cb84c-684a-48ec-aa24-03f5a7f218ff" (UID: "395cb84c-684a-48ec-aa24-03f5a7f218ff"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.016008 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "395cb84c-684a-48ec-aa24-03f5a7f218ff" (UID: "395cb84c-684a-48ec-aa24-03f5a7f218ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.049972 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.050237 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.050255 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.050264 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjd46\" (UniqueName: \"kubernetes.io/projected/395cb84c-684a-48ec-aa24-03f5a7f218ff-kube-api-access-kjd46\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.050349 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/395cb84c-684a-48ec-aa24-03f5a7f218ff-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.050395 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395cb84c-684a-48ec-aa24-03f5a7f218ff-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.056120 4726 generic.go:334] "Generic (PLEG): container finished" podID="395cb84c-684a-48ec-aa24-03f5a7f218ff" containerID="698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855" exitCode=2 Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.056153 4726 generic.go:334] "Generic (PLEG): container finished" podID="395cb84c-684a-48ec-aa24-03f5a7f218ff" containerID="7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5" exitCode=0 Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.056173 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.056178 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395cb84c-684a-48ec-aa24-03f5a7f218ff","Type":"ContainerDied","Data":"698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855"} Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.059245 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395cb84c-684a-48ec-aa24-03f5a7f218ff","Type":"ContainerDied","Data":"7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5"} Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.059271 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"395cb84c-684a-48ec-aa24-03f5a7f218ff","Type":"ContainerDied","Data":"9fef1c6274d546d30da9bf9a726cd36c9820448482b452f0163a55540d88fb55"} Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.059293 4726 scope.go:117] "RemoveContainer" containerID="698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.080322 4726 scope.go:117] "RemoveContainer" containerID="7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.111894 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.119219 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.126193 4726 scope.go:117] "RemoveContainer" containerID="698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855" Oct 01 16:17:46 crc kubenswrapper[4726]: E1001 16:17:46.127089 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855\": container with ID starting with 698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855 not found: ID does not exist" containerID="698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.127133 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855"} err="failed to get container status \"698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855\": rpc error: code = NotFound desc = could not find container \"698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855\": container with ID starting with 698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855 not found: ID does not exist" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.127157 4726 scope.go:117] "RemoveContainer" containerID="7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5" Oct 01 16:17:46 crc kubenswrapper[4726]: E1001 16:17:46.127498 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5\": container with ID starting with 7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5 not found: ID does not exist" containerID="7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.127528 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5"} err="failed to get container status \"7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5\": rpc error: code = NotFound desc = could not find container \"7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5\": container with ID starting with 7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5 not found: ID does not exist" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.127549 4726 scope.go:117] "RemoveContainer" containerID="698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.127777 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855"} err="failed to get container status \"698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855\": rpc error: code = NotFound desc = could not find container \"698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855\": container with ID starting with 698ec0a8372f4dcca061123f712acb59edf6d2362e1531b691c0c6ac26ee8855 not found: ID does not exist" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.127802 4726 scope.go:117] "RemoveContainer" containerID="7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.128010 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5"} err="failed to get container status \"7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5\": rpc error: code = NotFound desc = could not find container \"7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5\": container with ID starting with 7a204daff382f4bf1b8b64fcfab37de3090fe2236ae7333564547056729b96a5 not found: ID does not exist" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.138712 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:17:46 crc kubenswrapper[4726]: E1001 16:17:46.139110 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="395cb84c-684a-48ec-aa24-03f5a7f218ff" containerName="ceilometer-notification-agent" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.139127 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="395cb84c-684a-48ec-aa24-03f5a7f218ff" containerName="ceilometer-notification-agent" Oct 01 16:17:46 crc kubenswrapper[4726]: E1001 16:17:46.139153 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" containerName="init" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.139159 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" containerName="init" Oct 01 16:17:46 crc kubenswrapper[4726]: E1001 16:17:46.139170 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" containerName="dnsmasq-dns" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.139175 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" containerName="dnsmasq-dns" Oct 01 16:17:46 crc kubenswrapper[4726]: E1001 16:17:46.139195 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="395cb84c-684a-48ec-aa24-03f5a7f218ff" containerName="sg-core" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.139201 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="395cb84c-684a-48ec-aa24-03f5a7f218ff" containerName="sg-core" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.139346 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="395cb84c-684a-48ec-aa24-03f5a7f218ff" containerName="ceilometer-notification-agent" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.139366 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5c407e4-2f8c-47dd-85b4-b29c12ea3aed" containerName="dnsmasq-dns" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.139377 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="395cb84c-684a-48ec-aa24-03f5a7f218ff" containerName="sg-core" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.140839 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.148112 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.148266 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.157680 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.252946 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-config-data\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.252981 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/547c9bca-03ba-4d79-8c13-c8506ce27252-run-httpd\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.253027 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-scripts\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.253092 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w88sr\" (UniqueName: \"kubernetes.io/projected/547c9bca-03ba-4d79-8c13-c8506ce27252-kube-api-access-w88sr\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.253119 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/547c9bca-03ba-4d79-8c13-c8506ce27252-log-httpd\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.253150 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.253182 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.355006 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.355170 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-config-data\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.355196 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/547c9bca-03ba-4d79-8c13-c8506ce27252-run-httpd\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.355845 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/547c9bca-03ba-4d79-8c13-c8506ce27252-run-httpd\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.356093 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-scripts\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.356132 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w88sr\" (UniqueName: \"kubernetes.io/projected/547c9bca-03ba-4d79-8c13-c8506ce27252-kube-api-access-w88sr\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.356194 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/547c9bca-03ba-4d79-8c13-c8506ce27252-log-httpd\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.356241 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.356942 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/547c9bca-03ba-4d79-8c13-c8506ce27252-log-httpd\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.359501 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-scripts\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.360195 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-config-data\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.361639 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.364662 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.375507 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w88sr\" (UniqueName: \"kubernetes.io/projected/547c9bca-03ba-4d79-8c13-c8506ce27252-kube-api-access-w88sr\") pod \"ceilometer-0\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " pod="openstack/ceilometer-0" Oct 01 16:17:46 crc kubenswrapper[4726]: I1001 16:17:46.563596 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:17:47 crc kubenswrapper[4726]: I1001 16:17:47.067663 4726 generic.go:334] "Generic (PLEG): container finished" podID="5b0861ea-bb92-4592-9129-f9b3b4725a51" containerID="fd9d715e9acaeb8f62267eb263a5863c5e412082eab1db4177fd27caa439f20a" exitCode=0 Oct 01 16:17:47 crc kubenswrapper[4726]: I1001 16:17:47.067725 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6ccb6664-v52zf" event={"ID":"5b0861ea-bb92-4592-9129-f9b3b4725a51","Type":"ContainerDied","Data":"fd9d715e9acaeb8f62267eb263a5863c5e412082eab1db4177fd27caa439f20a"} Oct 01 16:17:47 crc kubenswrapper[4726]: W1001 16:17:47.110824 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod547c9bca_03ba_4d79_8c13_c8506ce27252.slice/crio-151710af01bedbf2ef9f8d74fe9aaa4e8df6eb5fb1f917c459a93cb1539f4978 WatchSource:0}: Error finding container 151710af01bedbf2ef9f8d74fe9aaa4e8df6eb5fb1f917c459a93cb1539f4978: Status 404 returned error can't find the container with id 151710af01bedbf2ef9f8d74fe9aaa4e8df6eb5fb1f917c459a93cb1539f4978 Oct 01 16:17:47 crc kubenswrapper[4726]: I1001 16:17:47.116681 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:17:47 crc kubenswrapper[4726]: I1001 16:17:47.824752 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="395cb84c-684a-48ec-aa24-03f5a7f218ff" path="/var/lib/kubelet/pods/395cb84c-684a-48ec-aa24-03f5a7f218ff/volumes" Oct 01 16:17:48 crc kubenswrapper[4726]: I1001 16:17:48.078698 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"547c9bca-03ba-4d79-8c13-c8506ce27252","Type":"ContainerStarted","Data":"151710af01bedbf2ef9f8d74fe9aaa4e8df6eb5fb1f917c459a93cb1539f4978"} Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.085161 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.090207 4726 generic.go:334] "Generic (PLEG): container finished" podID="4cff8692-89be-4a73-b714-1cc1e23a6b7d" containerID="258d4058f1e0c975c4b8b6d095a9ac22bfd17548d3b7a3a21debf9ab8df075dc" exitCode=137 Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.090239 4726 generic.go:334] "Generic (PLEG): container finished" podID="4cff8692-89be-4a73-b714-1cc1e23a6b7d" containerID="7e72b38cef22da8024b677ace27f67c37f8e798b5cd5134a36f6376bde1188f8" exitCode=137 Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.090288 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69b99cdcfc-z8t52" event={"ID":"4cff8692-89be-4a73-b714-1cc1e23a6b7d","Type":"ContainerDied","Data":"258d4058f1e0c975c4b8b6d095a9ac22bfd17548d3b7a3a21debf9ab8df075dc"} Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.090317 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69b99cdcfc-z8t52" event={"ID":"4cff8692-89be-4a73-b714-1cc1e23a6b7d","Type":"ContainerDied","Data":"7e72b38cef22da8024b677ace27f67c37f8e798b5cd5134a36f6376bde1188f8"} Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.092550 4726 generic.go:334] "Generic (PLEG): container finished" podID="a2911957-529c-46a4-8b63-e98beb36809b" containerID="ddda2e2ff2cccd5cef2be818e15e4aa658a852504f64ec8cabb4a25f0744877c" exitCode=137 Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.092589 4726 generic.go:334] "Generic (PLEG): container finished" podID="a2911957-529c-46a4-8b63-e98beb36809b" containerID="c0a9b35e55b27be633cb455b9f05dd6fdbf7f9323e7c364c5a34d99a677bf3cb" exitCode=137 Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.092638 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7945d7f85f-2qfqm" event={"ID":"a2911957-529c-46a4-8b63-e98beb36809b","Type":"ContainerDied","Data":"ddda2e2ff2cccd5cef2be818e15e4aa658a852504f64ec8cabb4a25f0744877c"} Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.092666 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7945d7f85f-2qfqm" event={"ID":"a2911957-529c-46a4-8b63-e98beb36809b","Type":"ContainerDied","Data":"c0a9b35e55b27be633cb455b9f05dd6fdbf7f9323e7c364c5a34d99a677bf3cb"} Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.094746 4726 generic.go:334] "Generic (PLEG): container finished" podID="62a37356-6f11-4c75-a452-570a847a0518" containerID="dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9" exitCode=137 Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.094776 4726 generic.go:334] "Generic (PLEG): container finished" podID="62a37356-6f11-4c75-a452-570a847a0518" containerID="d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d" exitCode=137 Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.094814 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fc7f8d489-j7grp" event={"ID":"62a37356-6f11-4c75-a452-570a847a0518","Type":"ContainerDied","Data":"dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9"} Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.094837 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fc7f8d489-j7grp" event={"ID":"62a37356-6f11-4c75-a452-570a847a0518","Type":"ContainerDied","Data":"d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d"} Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.094850 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fc7f8d489-j7grp" event={"ID":"62a37356-6f11-4c75-a452-570a847a0518","Type":"ContainerDied","Data":"27393fa6ac70e4f5c2bb8f970c76f5d94f054021da3e82838b4ec2c2798e7363"} Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.094868 4726 scope.go:117] "RemoveContainer" containerID="dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.094998 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fc7f8d489-j7grp" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.098034 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"547c9bca-03ba-4d79-8c13-c8506ce27252","Type":"ContainerStarted","Data":"4afa9e937c19dec8329428804da5e638a30d95fb07f42650d2de6a12ddec0eab"} Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.098101 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"547c9bca-03ba-4d79-8c13-c8506ce27252","Type":"ContainerStarted","Data":"5c2a0c28fb1998f178aec0b143da14133c36382949f778b4035b4bb686b5d029"} Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.223206 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62a37356-6f11-4c75-a452-570a847a0518-logs\") pod \"62a37356-6f11-4c75-a452-570a847a0518\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.223422 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62a37356-6f11-4c75-a452-570a847a0518-scripts\") pod \"62a37356-6f11-4c75-a452-570a847a0518\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.223585 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w22z\" (UniqueName: \"kubernetes.io/projected/62a37356-6f11-4c75-a452-570a847a0518-kube-api-access-2w22z\") pod \"62a37356-6f11-4c75-a452-570a847a0518\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.223674 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62a37356-6f11-4c75-a452-570a847a0518-config-data\") pod \"62a37356-6f11-4c75-a452-570a847a0518\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.223705 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/62a37356-6f11-4c75-a452-570a847a0518-horizon-secret-key\") pod \"62a37356-6f11-4c75-a452-570a847a0518\" (UID: \"62a37356-6f11-4c75-a452-570a847a0518\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.224499 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62a37356-6f11-4c75-a452-570a847a0518-logs" (OuterVolumeSpecName: "logs") pod "62a37356-6f11-4c75-a452-570a847a0518" (UID: "62a37356-6f11-4c75-a452-570a847a0518"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.228957 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62a37356-6f11-4c75-a452-570a847a0518-kube-api-access-2w22z" (OuterVolumeSpecName: "kube-api-access-2w22z") pod "62a37356-6f11-4c75-a452-570a847a0518" (UID: "62a37356-6f11-4c75-a452-570a847a0518"). InnerVolumeSpecName "kube-api-access-2w22z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.229014 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62a37356-6f11-4c75-a452-570a847a0518-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "62a37356-6f11-4c75-a452-570a847a0518" (UID: "62a37356-6f11-4c75-a452-570a847a0518"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.252127 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62a37356-6f11-4c75-a452-570a847a0518-scripts" (OuterVolumeSpecName: "scripts") pod "62a37356-6f11-4c75-a452-570a847a0518" (UID: "62a37356-6f11-4c75-a452-570a847a0518"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.254165 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62a37356-6f11-4c75-a452-570a847a0518-config-data" (OuterVolumeSpecName: "config-data") pod "62a37356-6f11-4c75-a452-570a847a0518" (UID: "62a37356-6f11-4c75-a452-570a847a0518"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.296691 4726 scope.go:117] "RemoveContainer" containerID="d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.311601 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.323328 4726 scope.go:117] "RemoveContainer" containerID="dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9" Oct 01 16:17:49 crc kubenswrapper[4726]: E1001 16:17:49.323720 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9\": container with ID starting with dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9 not found: ID does not exist" containerID="dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.323749 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9"} err="failed to get container status \"dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9\": rpc error: code = NotFound desc = could not find container \"dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9\": container with ID starting with dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9 not found: ID does not exist" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.323774 4726 scope.go:117] "RemoveContainer" containerID="d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d" Oct 01 16:17:49 crc kubenswrapper[4726]: E1001 16:17:49.324020 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d\": container with ID starting with d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d not found: ID does not exist" containerID="d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.324040 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d"} err="failed to get container status \"d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d\": rpc error: code = NotFound desc = could not find container \"d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d\": container with ID starting with d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d not found: ID does not exist" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.324068 4726 scope.go:117] "RemoveContainer" containerID="dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.324262 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9"} err="failed to get container status \"dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9\": rpc error: code = NotFound desc = could not find container \"dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9\": container with ID starting with dd4fcde80f27d80e6283255f2bfd0027e4c097e12e882bef3966e04a8bd6b8b9 not found: ID does not exist" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.324281 4726 scope.go:117] "RemoveContainer" containerID="d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.324442 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d"} err="failed to get container status \"d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d\": rpc error: code = NotFound desc = could not find container \"d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d\": container with ID starting with d4b56afa9f70ce467b5905beccf5f8a33cb61ee03da1239f1abf47e0d67ed23d not found: ID does not exist" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.325591 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62a37356-6f11-4c75-a452-570a847a0518-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.325618 4726 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/62a37356-6f11-4c75-a452-570a847a0518-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.325627 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62a37356-6f11-4c75-a452-570a847a0518-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.325637 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62a37356-6f11-4c75-a452-570a847a0518-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.325644 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w22z\" (UniqueName: \"kubernetes.io/projected/62a37356-6f11-4c75-a452-570a847a0518-kube-api-access-2w22z\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.427148 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cff8692-89be-4a73-b714-1cc1e23a6b7d-logs\") pod \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.428481 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4cff8692-89be-4a73-b714-1cc1e23a6b7d-horizon-secret-key\") pod \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.428572 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4cff8692-89be-4a73-b714-1cc1e23a6b7d-config-data\") pod \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.428616 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4cff8692-89be-4a73-b714-1cc1e23a6b7d-scripts\") pod \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.428634 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hknfw\" (UniqueName: \"kubernetes.io/projected/4cff8692-89be-4a73-b714-1cc1e23a6b7d-kube-api-access-hknfw\") pod \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\" (UID: \"4cff8692-89be-4a73-b714-1cc1e23a6b7d\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.428736 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cff8692-89be-4a73-b714-1cc1e23a6b7d-logs" (OuterVolumeSpecName: "logs") pod "4cff8692-89be-4a73-b714-1cc1e23a6b7d" (UID: "4cff8692-89be-4a73-b714-1cc1e23a6b7d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.429198 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cff8692-89be-4a73-b714-1cc1e23a6b7d-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.435246 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cff8692-89be-4a73-b714-1cc1e23a6b7d-kube-api-access-hknfw" (OuterVolumeSpecName: "kube-api-access-hknfw") pod "4cff8692-89be-4a73-b714-1cc1e23a6b7d" (UID: "4cff8692-89be-4a73-b714-1cc1e23a6b7d"). InnerVolumeSpecName "kube-api-access-hknfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.436411 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cff8692-89be-4a73-b714-1cc1e23a6b7d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "4cff8692-89be-4a73-b714-1cc1e23a6b7d" (UID: "4cff8692-89be-4a73-b714-1cc1e23a6b7d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.439690 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-fc7f8d489-j7grp"] Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.446558 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-fc7f8d489-j7grp"] Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.459481 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cff8692-89be-4a73-b714-1cc1e23a6b7d-scripts" (OuterVolumeSpecName: "scripts") pod "4cff8692-89be-4a73-b714-1cc1e23a6b7d" (UID: "4cff8692-89be-4a73-b714-1cc1e23a6b7d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.461165 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cff8692-89be-4a73-b714-1cc1e23a6b7d-config-data" (OuterVolumeSpecName: "config-data") pod "4cff8692-89be-4a73-b714-1cc1e23a6b7d" (UID: "4cff8692-89be-4a73-b714-1cc1e23a6b7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.476760 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.530910 4726 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4cff8692-89be-4a73-b714-1cc1e23a6b7d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.530943 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4cff8692-89be-4a73-b714-1cc1e23a6b7d-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.530954 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4cff8692-89be-4a73-b714-1cc1e23a6b7d-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.530962 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hknfw\" (UniqueName: \"kubernetes.io/projected/4cff8692-89be-4a73-b714-1cc1e23a6b7d-kube-api-access-hknfw\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.632444 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a2911957-529c-46a4-8b63-e98beb36809b-horizon-secret-key\") pod \"a2911957-529c-46a4-8b63-e98beb36809b\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.632816 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w5s9\" (UniqueName: \"kubernetes.io/projected/a2911957-529c-46a4-8b63-e98beb36809b-kube-api-access-7w5s9\") pod \"a2911957-529c-46a4-8b63-e98beb36809b\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.632985 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2911957-529c-46a4-8b63-e98beb36809b-logs\") pod \"a2911957-529c-46a4-8b63-e98beb36809b\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.633233 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a2911957-529c-46a4-8b63-e98beb36809b-scripts\") pod \"a2911957-529c-46a4-8b63-e98beb36809b\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.633450 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2911957-529c-46a4-8b63-e98beb36809b-config-data\") pod \"a2911957-529c-46a4-8b63-e98beb36809b\" (UID: \"a2911957-529c-46a4-8b63-e98beb36809b\") " Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.633536 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2911957-529c-46a4-8b63-e98beb36809b-logs" (OuterVolumeSpecName: "logs") pod "a2911957-529c-46a4-8b63-e98beb36809b" (UID: "a2911957-529c-46a4-8b63-e98beb36809b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.634078 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2911957-529c-46a4-8b63-e98beb36809b-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.635578 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2911957-529c-46a4-8b63-e98beb36809b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a2911957-529c-46a4-8b63-e98beb36809b" (UID: "a2911957-529c-46a4-8b63-e98beb36809b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.637191 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2911957-529c-46a4-8b63-e98beb36809b-kube-api-access-7w5s9" (OuterVolumeSpecName: "kube-api-access-7w5s9") pod "a2911957-529c-46a4-8b63-e98beb36809b" (UID: "a2911957-529c-46a4-8b63-e98beb36809b"). InnerVolumeSpecName "kube-api-access-7w5s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.658101 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2911957-529c-46a4-8b63-e98beb36809b-scripts" (OuterVolumeSpecName: "scripts") pod "a2911957-529c-46a4-8b63-e98beb36809b" (UID: "a2911957-529c-46a4-8b63-e98beb36809b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.661705 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2911957-529c-46a4-8b63-e98beb36809b-config-data" (OuterVolumeSpecName: "config-data") pod "a2911957-529c-46a4-8b63-e98beb36809b" (UID: "a2911957-529c-46a4-8b63-e98beb36809b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.736626 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w5s9\" (UniqueName: \"kubernetes.io/projected/a2911957-529c-46a4-8b63-e98beb36809b-kube-api-access-7w5s9\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.736672 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a2911957-529c-46a4-8b63-e98beb36809b-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.736691 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2911957-529c-46a4-8b63-e98beb36809b-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.736710 4726 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a2911957-529c-46a4-8b63-e98beb36809b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:49 crc kubenswrapper[4726]: I1001 16:17:49.826543 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62a37356-6f11-4c75-a452-570a847a0518" path="/var/lib/kubelet/pods/62a37356-6f11-4c75-a452-570a847a0518/volumes" Oct 01 16:17:50 crc kubenswrapper[4726]: I1001 16:17:50.113526 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"547c9bca-03ba-4d79-8c13-c8506ce27252","Type":"ContainerStarted","Data":"73c80c2c9d0d8cfaa3cca243004a745910379c837d85a4e6bce907ed5e68bf1c"} Oct 01 16:17:50 crc kubenswrapper[4726]: I1001 16:17:50.119496 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69b99cdcfc-z8t52" event={"ID":"4cff8692-89be-4a73-b714-1cc1e23a6b7d","Type":"ContainerDied","Data":"e36f6ec5d434f48d3fc32b864c1d3c45c7a2d4e6614a5b7e2158dca1a2417af8"} Oct 01 16:17:50 crc kubenswrapper[4726]: I1001 16:17:50.119609 4726 scope.go:117] "RemoveContainer" containerID="258d4058f1e0c975c4b8b6d095a9ac22bfd17548d3b7a3a21debf9ab8df075dc" Oct 01 16:17:50 crc kubenswrapper[4726]: I1001 16:17:50.119618 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69b99cdcfc-z8t52" Oct 01 16:17:50 crc kubenswrapper[4726]: I1001 16:17:50.123409 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7945d7f85f-2qfqm" Oct 01 16:17:50 crc kubenswrapper[4726]: I1001 16:17:50.123446 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7945d7f85f-2qfqm" event={"ID":"a2911957-529c-46a4-8b63-e98beb36809b","Type":"ContainerDied","Data":"a78c36cc993a32aa321505fadae8d604c2cf99733fb6eeaacf292b46c3cf925b"} Oct 01 16:17:50 crc kubenswrapper[4726]: I1001 16:17:50.168268 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-69b99cdcfc-z8t52"] Oct 01 16:17:50 crc kubenswrapper[4726]: I1001 16:17:50.177485 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-69b99cdcfc-z8t52"] Oct 01 16:17:50 crc kubenswrapper[4726]: I1001 16:17:50.185666 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7945d7f85f-2qfqm"] Oct 01 16:17:50 crc kubenswrapper[4726]: I1001 16:17:50.195761 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7945d7f85f-2qfqm"] Oct 01 16:17:50 crc kubenswrapper[4726]: I1001 16:17:50.329528 4726 scope.go:117] "RemoveContainer" containerID="7e72b38cef22da8024b677ace27f67c37f8e798b5cd5134a36f6376bde1188f8" Oct 01 16:17:50 crc kubenswrapper[4726]: I1001 16:17:50.358740 4726 scope.go:117] "RemoveContainer" containerID="ddda2e2ff2cccd5cef2be818e15e4aa658a852504f64ec8cabb4a25f0744877c" Oct 01 16:17:50 crc kubenswrapper[4726]: I1001 16:17:50.549760 4726 scope.go:117] "RemoveContainer" containerID="c0a9b35e55b27be633cb455b9f05dd6fdbf7f9323e7c364c5a34d99a677bf3cb" Oct 01 16:17:51 crc kubenswrapper[4726]: I1001 16:17:51.822040 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cff8692-89be-4a73-b714-1cc1e23a6b7d" path="/var/lib/kubelet/pods/4cff8692-89be-4a73-b714-1cc1e23a6b7d/volumes" Oct 01 16:17:51 crc kubenswrapper[4726]: I1001 16:17:51.823309 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2911957-529c-46a4-8b63-e98beb36809b" path="/var/lib/kubelet/pods/a2911957-529c-46a4-8b63-e98beb36809b/volumes" Oct 01 16:17:52 crc kubenswrapper[4726]: I1001 16:17:52.159396 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"547c9bca-03ba-4d79-8c13-c8506ce27252","Type":"ContainerStarted","Data":"afd47f4eca931d1883629f439f47d59026063730435b0d205c5e545731a91049"} Oct 01 16:17:52 crc kubenswrapper[4726]: I1001 16:17:52.160355 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 16:17:52 crc kubenswrapper[4726]: I1001 16:17:52.178819 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.7709346080000001 podStartE2EDuration="6.178801784s" podCreationTimestamp="2025-10-01 16:17:46 +0000 UTC" firstStartedPulling="2025-10-01 16:17:47.113120517 +0000 UTC m=+1120.014673094" lastFinishedPulling="2025-10-01 16:17:51.520987683 +0000 UTC m=+1124.422540270" observedRunningTime="2025-10-01 16:17:52.177702303 +0000 UTC m=+1125.079254900" watchObservedRunningTime="2025-10-01 16:17:52.178801784 +0000 UTC m=+1125.080354361" Oct 01 16:17:53 crc kubenswrapper[4726]: I1001 16:17:53.413683 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:17:53 crc kubenswrapper[4726]: I1001 16:17:53.414042 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:17:54 crc kubenswrapper[4726]: I1001 16:17:54.179862 4726 generic.go:334] "Generic (PLEG): container finished" podID="fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6" containerID="0fd712d7373def2d15480febabcafaa3ba5119977c938919054b21fd5ef60748" exitCode=0 Oct 01 16:17:54 crc kubenswrapper[4726]: I1001 16:17:54.179936 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bhxd2" event={"ID":"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6","Type":"ContainerDied","Data":"0fd712d7373def2d15480febabcafaa3ba5119977c938919054b21fd5ef60748"} Oct 01 16:17:55 crc kubenswrapper[4726]: E1001 16:17:55.156638 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda83273b9_e1f0_47e6_8bbe_839f4c0075b3.slice/crio-conmon-16d3ded60afe96d427098a307e71f20d4f35f12993be114dad1f6c2582f663a1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda83273b9_e1f0_47e6_8bbe_839f4c0075b3.slice/crio-16d3ded60afe96d427098a307e71f20d4f35f12993be114dad1f6c2582f663a1.scope\": RecentStats: unable to find data in memory cache]" Oct 01 16:17:55 crc kubenswrapper[4726]: I1001 16:17:55.192822 4726 generic.go:334] "Generic (PLEG): container finished" podID="a83273b9-e1f0-47e6-8bbe-839f4c0075b3" containerID="16d3ded60afe96d427098a307e71f20d4f35f12993be114dad1f6c2582f663a1" exitCode=0 Oct 01 16:17:55 crc kubenswrapper[4726]: I1001 16:17:55.192901 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gsljc" event={"ID":"a83273b9-e1f0-47e6-8bbe-839f4c0075b3","Type":"ContainerDied","Data":"16d3ded60afe96d427098a307e71f20d4f35f12993be114dad1f6c2582f663a1"} Oct 01 16:17:55 crc kubenswrapper[4726]: I1001 16:17:55.492432 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bhxd2" Oct 01 16:17:55 crc kubenswrapper[4726]: I1001 16:17:55.563643 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-db-sync-config-data\") pod \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\" (UID: \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\") " Oct 01 16:17:55 crc kubenswrapper[4726]: I1001 16:17:55.563914 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hd4s2\" (UniqueName: \"kubernetes.io/projected/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-kube-api-access-hd4s2\") pod \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\" (UID: \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\") " Oct 01 16:17:55 crc kubenswrapper[4726]: I1001 16:17:55.564113 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-combined-ca-bundle\") pod \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\" (UID: \"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6\") " Oct 01 16:17:55 crc kubenswrapper[4726]: I1001 16:17:55.572014 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6" (UID: "fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:55 crc kubenswrapper[4726]: I1001 16:17:55.572029 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-kube-api-access-hd4s2" (OuterVolumeSpecName: "kube-api-access-hd4s2") pod "fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6" (UID: "fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6"). InnerVolumeSpecName "kube-api-access-hd4s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:55 crc kubenswrapper[4726]: I1001 16:17:55.595559 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6" (UID: "fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:55 crc kubenswrapper[4726]: I1001 16:17:55.667040 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hd4s2\" (UniqueName: \"kubernetes.io/projected/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-kube-api-access-hd4s2\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:55 crc kubenswrapper[4726]: I1001 16:17:55.667106 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:55 crc kubenswrapper[4726]: I1001 16:17:55.667122 4726 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.206002 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bhxd2" event={"ID":"fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6","Type":"ContainerDied","Data":"f57d82a6060294c1b43e8117ffc939abf7245e8f1ff27b262bcff2a4e8093ecb"} Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.206293 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f57d82a6060294c1b43e8117ffc939abf7245e8f1ff27b262bcff2a4e8093ecb" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.206340 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bhxd2" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.210503 4726 generic.go:334] "Generic (PLEG): container finished" podID="7b994ac8-f28c-458a-873f-4270e9f0f36b" containerID="79416f7542f476fdc76e78a15e310f7390f6670c401eb02b3b186018905708d4" exitCode=0 Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.210568 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-skmf7" event={"ID":"7b994ac8-f28c-458a-873f-4270e9f0f36b","Type":"ContainerDied","Data":"79416f7542f476fdc76e78a15e310f7390f6670c401eb02b3b186018905708d4"} Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.479908 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-57cbb4948c-gcc56"] Oct 01 16:17:56 crc kubenswrapper[4726]: E1001 16:17:56.480360 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62a37356-6f11-4c75-a452-570a847a0518" containerName="horizon" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.480385 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="62a37356-6f11-4c75-a452-570a847a0518" containerName="horizon" Oct 01 16:17:56 crc kubenswrapper[4726]: E1001 16:17:56.480404 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62a37356-6f11-4c75-a452-570a847a0518" containerName="horizon-log" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.480412 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="62a37356-6f11-4c75-a452-570a847a0518" containerName="horizon-log" Oct 01 16:17:56 crc kubenswrapper[4726]: E1001 16:17:56.480438 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2911957-529c-46a4-8b63-e98beb36809b" containerName="horizon" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.480446 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2911957-529c-46a4-8b63-e98beb36809b" containerName="horizon" Oct 01 16:17:56 crc kubenswrapper[4726]: E1001 16:17:56.480460 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cff8692-89be-4a73-b714-1cc1e23a6b7d" containerName="horizon" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.480467 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cff8692-89be-4a73-b714-1cc1e23a6b7d" containerName="horizon" Oct 01 16:17:56 crc kubenswrapper[4726]: E1001 16:17:56.480484 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6" containerName="barbican-db-sync" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.480492 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6" containerName="barbican-db-sync" Oct 01 16:17:56 crc kubenswrapper[4726]: E1001 16:17:56.497256 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2911957-529c-46a4-8b63-e98beb36809b" containerName="horizon-log" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.497297 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2911957-529c-46a4-8b63-e98beb36809b" containerName="horizon-log" Oct 01 16:17:56 crc kubenswrapper[4726]: E1001 16:17:56.497320 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cff8692-89be-4a73-b714-1cc1e23a6b7d" containerName="horizon-log" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.497327 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cff8692-89be-4a73-b714-1cc1e23a6b7d" containerName="horizon-log" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.497665 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="62a37356-6f11-4c75-a452-570a847a0518" containerName="horizon-log" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.497698 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cff8692-89be-4a73-b714-1cc1e23a6b7d" containerName="horizon" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.497710 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2911957-529c-46a4-8b63-e98beb36809b" containerName="horizon-log" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.497727 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2911957-529c-46a4-8b63-e98beb36809b" containerName="horizon" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.497752 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="62a37356-6f11-4c75-a452-570a847a0518" containerName="horizon" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.497763 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cff8692-89be-4a73-b714-1cc1e23a6b7d" containerName="horizon-log" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.497778 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6" containerName="barbican-db-sync" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.498853 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-99545b666-4wmpl"] Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.499255 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.501161 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.510242 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-57cbb4948c-gcc56"] Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.511383 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.511712 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.511825 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-f2dgd" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.512313 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.523539 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-99545b666-4wmpl"] Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.583807 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f25ba81-3f98-4d63-8035-2b30a1df5f54-config-data-custom\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.583856 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94005d17-9bb0-4039-b493-5057a554f054-config-data\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.583917 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94005d17-9bb0-4039-b493-5057a554f054-combined-ca-bundle\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.583937 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f25ba81-3f98-4d63-8035-2b30a1df5f54-logs\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.583967 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94005d17-9bb0-4039-b493-5057a554f054-logs\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.583987 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f25ba81-3f98-4d63-8035-2b30a1df5f54-combined-ca-bundle\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.584015 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94005d17-9bb0-4039-b493-5057a554f054-config-data-custom\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.584039 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f25ba81-3f98-4d63-8035-2b30a1df5f54-config-data\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.584077 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf99z\" (UniqueName: \"kubernetes.io/projected/7f25ba81-3f98-4d63-8035-2b30a1df5f54-kube-api-access-hf99z\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.584100 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg4vv\" (UniqueName: \"kubernetes.io/projected/94005d17-9bb0-4039-b493-5057a554f054-kube-api-access-pg4vv\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.593028 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gsljc" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.648605 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-gk7lw"] Oct 01 16:17:56 crc kubenswrapper[4726]: E1001 16:17:56.649341 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a83273b9-e1f0-47e6-8bbe-839f4c0075b3" containerName="cinder-db-sync" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.649460 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a83273b9-e1f0-47e6-8bbe-839f4c0075b3" containerName="cinder-db-sync" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.649777 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a83273b9-e1f0-47e6-8bbe-839f4c0075b3" containerName="cinder-db-sync" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.651107 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.654779 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-gk7lw"] Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.684708 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-scripts\") pod \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.684750 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-config-data\") pod \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.684819 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-combined-ca-bundle\") pod \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.684862 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-etc-machine-id\") pod \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.684913 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znxm2\" (UniqueName: \"kubernetes.io/projected/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-kube-api-access-znxm2\") pod \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.684948 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-db-sync-config-data\") pod \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\" (UID: \"a83273b9-e1f0-47e6-8bbe-839f4c0075b3\") " Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.685340 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94005d17-9bb0-4039-b493-5057a554f054-combined-ca-bundle\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.685368 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f25ba81-3f98-4d63-8035-2b30a1df5f54-logs\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.685394 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94005d17-9bb0-4039-b493-5057a554f054-logs\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.685414 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f25ba81-3f98-4d63-8035-2b30a1df5f54-combined-ca-bundle\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.685448 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94005d17-9bb0-4039-b493-5057a554f054-config-data-custom\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.685469 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f25ba81-3f98-4d63-8035-2b30a1df5f54-config-data\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.685497 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf99z\" (UniqueName: \"kubernetes.io/projected/7f25ba81-3f98-4d63-8035-2b30a1df5f54-kube-api-access-hf99z\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.685516 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg4vv\" (UniqueName: \"kubernetes.io/projected/94005d17-9bb0-4039-b493-5057a554f054-kube-api-access-pg4vv\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.685557 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f25ba81-3f98-4d63-8035-2b30a1df5f54-config-data-custom\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.685580 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94005d17-9bb0-4039-b493-5057a554f054-config-data\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.687795 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94005d17-9bb0-4039-b493-5057a554f054-logs\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.691926 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a83273b9-e1f0-47e6-8bbe-839f4c0075b3" (UID: "a83273b9-e1f0-47e6-8bbe-839f4c0075b3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.691946 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a83273b9-e1f0-47e6-8bbe-839f4c0075b3" (UID: "a83273b9-e1f0-47e6-8bbe-839f4c0075b3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.691936 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f25ba81-3f98-4d63-8035-2b30a1df5f54-logs\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.694978 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94005d17-9bb0-4039-b493-5057a554f054-config-data-custom\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.695370 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f25ba81-3f98-4d63-8035-2b30a1df5f54-config-data\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.696547 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f25ba81-3f98-4d63-8035-2b30a1df5f54-config-data-custom\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.700697 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94005d17-9bb0-4039-b493-5057a554f054-config-data\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.702714 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-kube-api-access-znxm2" (OuterVolumeSpecName: "kube-api-access-znxm2") pod "a83273b9-e1f0-47e6-8bbe-839f4c0075b3" (UID: "a83273b9-e1f0-47e6-8bbe-839f4c0075b3"). InnerVolumeSpecName "kube-api-access-znxm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.708791 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-scripts" (OuterVolumeSpecName: "scripts") pod "a83273b9-e1f0-47e6-8bbe-839f4c0075b3" (UID: "a83273b9-e1f0-47e6-8bbe-839f4c0075b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.709709 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg4vv\" (UniqueName: \"kubernetes.io/projected/94005d17-9bb0-4039-b493-5057a554f054-kube-api-access-pg4vv\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.713862 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94005d17-9bb0-4039-b493-5057a554f054-combined-ca-bundle\") pod \"barbican-worker-57cbb4948c-gcc56\" (UID: \"94005d17-9bb0-4039-b493-5057a554f054\") " pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.716416 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f25ba81-3f98-4d63-8035-2b30a1df5f54-combined-ca-bundle\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.719376 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a83273b9-e1f0-47e6-8bbe-839f4c0075b3" (UID: "a83273b9-e1f0-47e6-8bbe-839f4c0075b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.722666 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf99z\" (UniqueName: \"kubernetes.io/projected/7f25ba81-3f98-4d63-8035-2b30a1df5f54-kube-api-access-hf99z\") pod \"barbican-keystone-listener-99545b666-4wmpl\" (UID: \"7f25ba81-3f98-4d63-8035-2b30a1df5f54\") " pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.769128 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7bc64455bb-r2zhn"] Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.770935 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.773281 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-config-data" (OuterVolumeSpecName: "config-data") pod "a83273b9-e1f0-47e6-8bbe-839f4c0075b3" (UID: "a83273b9-e1f0-47e6-8bbe-839f4c0075b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.775790 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.782219 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7bc64455bb-r2zhn"] Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.788810 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.788879 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnrnq\" (UniqueName: \"kubernetes.io/projected/b096e2a5-08e3-4aa6-b98c-50101f0229d9-kube-api-access-cnrnq\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.788958 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-config\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.789004 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.789033 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.789108 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.789156 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.789167 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.789176 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.789186 4726 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.789194 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znxm2\" (UniqueName: \"kubernetes.io/projected/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-kube-api-access-znxm2\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.789204 4726 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a83273b9-e1f0-47e6-8bbe-839f4c0075b3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.891129 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.891267 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-config-data\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.891304 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-combined-ca-bundle\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.891335 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.891368 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnrnq\" (UniqueName: \"kubernetes.io/projected/b096e2a5-08e3-4aa6-b98c-50101f0229d9-kube-api-access-cnrnq\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.891396 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fh54\" (UniqueName: \"kubernetes.io/projected/ebde1298-9550-4cac-a7a4-a4f31135b59f-kube-api-access-6fh54\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.891431 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebde1298-9550-4cac-a7a4-a4f31135b59f-logs\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.891509 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-config-data-custom\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.891539 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-config\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.891581 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.891609 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.892296 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.892558 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.893019 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.894331 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.894375 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-config\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.911462 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnrnq\" (UniqueName: \"kubernetes.io/projected/b096e2a5-08e3-4aa6-b98c-50101f0229d9-kube-api-access-cnrnq\") pod \"dnsmasq-dns-7c67bffd47-gk7lw\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.924914 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-57cbb4948c-gcc56" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.948666 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-99545b666-4wmpl" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.976224 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.993220 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fh54\" (UniqueName: \"kubernetes.io/projected/ebde1298-9550-4cac-a7a4-a4f31135b59f-kube-api-access-6fh54\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.993269 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebde1298-9550-4cac-a7a4-a4f31135b59f-logs\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.993315 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-config-data-custom\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.993407 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-config-data\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.993434 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-combined-ca-bundle\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.994387 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebde1298-9550-4cac-a7a4-a4f31135b59f-logs\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.998934 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-combined-ca-bundle\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.999313 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-config-data\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:56 crc kubenswrapper[4726]: I1001 16:17:56.999574 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-config-data-custom\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.016126 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fh54\" (UniqueName: \"kubernetes.io/projected/ebde1298-9550-4cac-a7a4-a4f31135b59f-kube-api-access-6fh54\") pod \"barbican-api-7bc64455bb-r2zhn\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.146342 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.226157 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gsljc" event={"ID":"a83273b9-e1f0-47e6-8bbe-839f4c0075b3","Type":"ContainerDied","Data":"bb7c90e518c145514f1782850e7d523b3c628c6879372caf741bd9a193272360"} Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.226210 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gsljc" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.226231 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb7c90e518c145514f1782850e7d523b3c628c6879372caf741bd9a193272360" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.405931 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-57cbb4948c-gcc56"] Oct 01 16:17:57 crc kubenswrapper[4726]: W1001 16:17:57.436226 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94005d17_9bb0_4039_b493_5057a554f054.slice/crio-1a80d83384f3683cb995204d8bed48a5f7fd05fe42ed98bc01dd8d0e6e47efb4 WatchSource:0}: Error finding container 1a80d83384f3683cb995204d8bed48a5f7fd05fe42ed98bc01dd8d0e6e47efb4: Status 404 returned error can't find the container with id 1a80d83384f3683cb995204d8bed48a5f7fd05fe42ed98bc01dd8d0e6e47efb4 Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.539679 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-gk7lw"] Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.562194 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-99545b666-4wmpl"] Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.580085 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-gk7lw"] Oct 01 16:17:57 crc kubenswrapper[4726]: W1001 16:17:57.600528 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb096e2a5_08e3_4aa6_b98c_50101f0229d9.slice/crio-badd1da4f9256372804f64e41caaccedc52c06362a8e26b7a2ff2b6327ace719 WatchSource:0}: Error finding container badd1da4f9256372804f64e41caaccedc52c06362a8e26b7a2ff2b6327ace719: Status 404 returned error can't find the container with id badd1da4f9256372804f64e41caaccedc52c06362a8e26b7a2ff2b6327ace719 Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.645647 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.648470 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.651749 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-x9fkn" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.657951 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.660599 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.661100 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.703642 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.726530 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.727735 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx4z8\" (UniqueName: \"kubernetes.io/projected/e86b00b2-1546-4cfc-a51a-23b92e950bc8-kube-api-access-xx4z8\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.727906 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-config-data\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.728694 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e86b00b2-1546-4cfc-a51a-23b92e950bc8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.728772 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-scripts\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.728930 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.758230 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf"] Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.760025 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.770120 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf"] Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.779654 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-skmf7" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.799156 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:17:57 crc kubenswrapper[4726]: E1001 16:17:57.799647 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b994ac8-f28c-458a-873f-4270e9f0f36b" containerName="neutron-db-sync" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.799671 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b994ac8-f28c-458a-873f-4270e9f0f36b" containerName="neutron-db-sync" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.799883 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b994ac8-f28c-458a-873f-4270e9f0f36b" containerName="neutron-db-sync" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.804391 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.807217 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.832973 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-config-data\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.833039 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-config\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.833108 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e86b00b2-1546-4cfc-a51a-23b92e950bc8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.833126 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.833149 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-scripts\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.833190 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzmw9\" (UniqueName: \"kubernetes.io/projected/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-kube-api-access-gzmw9\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.833214 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.833245 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.833280 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.833300 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-dns-svc\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.833322 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.833344 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx4z8\" (UniqueName: \"kubernetes.io/projected/e86b00b2-1546-4cfc-a51a-23b92e950bc8-kube-api-access-xx4z8\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.834449 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e86b00b2-1546-4cfc-a51a-23b92e950bc8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.846275 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-config-data\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.848233 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.850220 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.860653 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-scripts\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.873715 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7bc64455bb-r2zhn"] Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.873793 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.874778 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx4z8\" (UniqueName: \"kubernetes.io/projected/e86b00b2-1546-4cfc-a51a-23b92e950bc8-kube-api-access-xx4z8\") pod \"cinder-scheduler-0\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.934642 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q72v4\" (UniqueName: \"kubernetes.io/projected/7b994ac8-f28c-458a-873f-4270e9f0f36b-kube-api-access-q72v4\") pod \"7b994ac8-f28c-458a-873f-4270e9f0f36b\" (UID: \"7b994ac8-f28c-458a-873f-4270e9f0f36b\") " Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.934841 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b994ac8-f28c-458a-873f-4270e9f0f36b-combined-ca-bundle\") pod \"7b994ac8-f28c-458a-873f-4270e9f0f36b\" (UID: \"7b994ac8-f28c-458a-873f-4270e9f0f36b\") " Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.934875 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7b994ac8-f28c-458a-873f-4270e9f0f36b-config\") pod \"7b994ac8-f28c-458a-873f-4270e9f0f36b\" (UID: \"7b994ac8-f28c-458a-873f-4270e9f0f36b\") " Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.935235 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.935275 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-config-data\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.935320 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-scripts\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.935350 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.935373 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-config-data-custom\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.935401 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-dns-svc\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.935477 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-config\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.935511 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c580861-1c2f-4f6b-933a-5203bd970f8a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.935532 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c580861-1c2f-4f6b-933a-5203bd970f8a-logs\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.935559 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.935586 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p7lx\" (UniqueName: \"kubernetes.io/projected/3c580861-1c2f-4f6b-933a-5203bd970f8a-kube-api-access-4p7lx\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.935622 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.935644 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzmw9\" (UniqueName: \"kubernetes.io/projected/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-kube-api-access-gzmw9\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.937527 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.937570 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-config\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.938288 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.938302 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.938921 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-dns-svc\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.956896 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b994ac8-f28c-458a-873f-4270e9f0f36b-kube-api-access-q72v4" (OuterVolumeSpecName: "kube-api-access-q72v4") pod "7b994ac8-f28c-458a-873f-4270e9f0f36b" (UID: "7b994ac8-f28c-458a-873f-4270e9f0f36b"). InnerVolumeSpecName "kube-api-access-q72v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.962968 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzmw9\" (UniqueName: \"kubernetes.io/projected/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-kube-api-access-gzmw9\") pod \"dnsmasq-dns-5cc8b5d5c5-lw9xf\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.998474 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b994ac8-f28c-458a-873f-4270e9f0f36b-config" (OuterVolumeSpecName: "config") pod "7b994ac8-f28c-458a-873f-4270e9f0f36b" (UID: "7b994ac8-f28c-458a-873f-4270e9f0f36b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:57 crc kubenswrapper[4726]: I1001 16:17:57.999280 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b994ac8-f28c-458a-873f-4270e9f0f36b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b994ac8-f28c-458a-873f-4270e9f0f36b" (UID: "7b994ac8-f28c-458a-873f-4270e9f0f36b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.037142 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c580861-1c2f-4f6b-933a-5203bd970f8a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.037182 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c580861-1c2f-4f6b-933a-5203bd970f8a-logs\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.037224 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p7lx\" (UniqueName: \"kubernetes.io/projected/3c580861-1c2f-4f6b-933a-5203bd970f8a-kube-api-access-4p7lx\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.037254 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.037277 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c580861-1c2f-4f6b-933a-5203bd970f8a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.037303 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-config-data\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.037389 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-scripts\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.037432 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-config-data-custom\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.037675 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q72v4\" (UniqueName: \"kubernetes.io/projected/7b994ac8-f28c-458a-873f-4270e9f0f36b-kube-api-access-q72v4\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.037690 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b994ac8-f28c-458a-873f-4270e9f0f36b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.037699 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/7b994ac8-f28c-458a-873f-4270e9f0f36b-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.039181 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c580861-1c2f-4f6b-933a-5203bd970f8a-logs\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.041710 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-scripts\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.042505 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.043263 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-config-data\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.043339 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-config-data-custom\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.057314 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p7lx\" (UniqueName: \"kubernetes.io/projected/3c580861-1c2f-4f6b-933a-5203bd970f8a-kube-api-access-4p7lx\") pod \"cinder-api-0\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.073354 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.126120 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.163143 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.238847 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57cbb4948c-gcc56" event={"ID":"94005d17-9bb0-4039-b493-5057a554f054","Type":"ContainerStarted","Data":"1a80d83384f3683cb995204d8bed48a5f7fd05fe42ed98bc01dd8d0e6e47efb4"} Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.240317 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc64455bb-r2zhn" event={"ID":"ebde1298-9550-4cac-a7a4-a4f31135b59f","Type":"ContainerStarted","Data":"83fe40dbcf64c59c5ce1fbc0b8a6f2e82f3e7e48ddf8edd837c82f91d78058e1"} Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.240336 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc64455bb-r2zhn" event={"ID":"ebde1298-9550-4cac-a7a4-a4f31135b59f","Type":"ContainerStarted","Data":"3fa7b320db45fb191c5c961a07f7e980b14bc8e1a564ce922d490fdea9e02b69"} Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.263526 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-99545b666-4wmpl" event={"ID":"7f25ba81-3f98-4d63-8035-2b30a1df5f54","Type":"ContainerStarted","Data":"ba07a54e6068ce184a1f476ce5bbf478f3843cf106af574b83c222da30024a0a"} Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.271987 4726 generic.go:334] "Generic (PLEG): container finished" podID="b096e2a5-08e3-4aa6-b98c-50101f0229d9" containerID="2d5a1ac6ed14d039f09b5804239e2481024122e6b4f674d144d55d452b18e175" exitCode=0 Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.272041 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" event={"ID":"b096e2a5-08e3-4aa6-b98c-50101f0229d9","Type":"ContainerDied","Data":"2d5a1ac6ed14d039f09b5804239e2481024122e6b4f674d144d55d452b18e175"} Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.272078 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" event={"ID":"b096e2a5-08e3-4aa6-b98c-50101f0229d9","Type":"ContainerStarted","Data":"badd1da4f9256372804f64e41caaccedc52c06362a8e26b7a2ff2b6327ace719"} Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.280297 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-skmf7" event={"ID":"7b994ac8-f28c-458a-873f-4270e9f0f36b","Type":"ContainerDied","Data":"7d3c1db43ffc80652e318211f6b2db23b9f65f3c63eae25c624bb98a6a8f3a8b"} Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.280347 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d3c1db43ffc80652e318211f6b2db23b9f65f3c63eae25c624bb98a6a8f3a8b" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.280415 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-skmf7" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.520723 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf"] Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.557108 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-czbjh"] Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.572670 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.591268 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-czbjh"] Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.635687 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.656764 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-config\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.657178 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.657313 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qf7r\" (UniqueName: \"kubernetes.io/projected/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-kube-api-access-2qf7r\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.657406 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.657544 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.657656 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-dns-svc\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.657884 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-566c76dfb4-x49km"] Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.659846 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.669185 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.669546 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-pdhpb" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.670115 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.670227 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.693869 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-566c76dfb4-x49km"] Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.745239 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf"] Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.759240 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-config\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.759336 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.759381 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qf7r\" (UniqueName: \"kubernetes.io/projected/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-kube-api-access-2qf7r\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.759411 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-config\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.759441 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.759469 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-ovndb-tls-certs\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.759495 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-httpd-config\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.759512 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4rjk\" (UniqueName: \"kubernetes.io/projected/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-kube-api-access-t4rjk\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.759531 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-combined-ca-bundle\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.759556 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.759573 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-dns-svc\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.760537 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-dns-svc\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.761192 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-config\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.761847 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.765093 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.765781 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.833646 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qf7r\" (UniqueName: \"kubernetes.io/projected/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-kube-api-access-2qf7r\") pod \"dnsmasq-dns-6578955fd5-czbjh\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.862250 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-config\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.862309 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-ovndb-tls-certs\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.862335 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-httpd-config\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.862355 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4rjk\" (UniqueName: \"kubernetes.io/projected/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-kube-api-access-t4rjk\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.862373 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-combined-ca-bundle\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.920072 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-ovndb-tls-certs\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.920965 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-combined-ca-bundle\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.931528 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-httpd-config\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.945202 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-config\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.949234 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.956316 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.963376 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-ovsdbserver-sb\") pod \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.963691 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnrnq\" (UniqueName: \"kubernetes.io/projected/b096e2a5-08e3-4aa6-b98c-50101f0229d9-kube-api-access-cnrnq\") pod \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.966765 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4rjk\" (UniqueName: \"kubernetes.io/projected/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-kube-api-access-t4rjk\") pod \"neutron-566c76dfb4-x49km\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:58 crc kubenswrapper[4726]: I1001 16:17:58.984079 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b096e2a5-08e3-4aa6-b98c-50101f0229d9-kube-api-access-cnrnq" (OuterVolumeSpecName: "kube-api-access-cnrnq") pod "b096e2a5-08e3-4aa6-b98c-50101f0229d9" (UID: "b096e2a5-08e3-4aa6-b98c-50101f0229d9"). InnerVolumeSpecName "kube-api-access-cnrnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.010503 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b096e2a5-08e3-4aa6-b98c-50101f0229d9" (UID: "b096e2a5-08e3-4aa6-b98c-50101f0229d9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.012446 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.025426 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.078747 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-dns-svc\") pod \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.078847 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-ovsdbserver-nb\") pod \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.078942 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-dns-swift-storage-0\") pod \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.078976 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-config\") pod \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\" (UID: \"b096e2a5-08e3-4aa6-b98c-50101f0229d9\") " Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.079334 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnrnq\" (UniqueName: \"kubernetes.io/projected/b096e2a5-08e3-4aa6-b98c-50101f0229d9-kube-api-access-cnrnq\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.079349 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.143851 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-config" (OuterVolumeSpecName: "config") pod "b096e2a5-08e3-4aa6-b98c-50101f0229d9" (UID: "b096e2a5-08e3-4aa6-b98c-50101f0229d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.162836 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b096e2a5-08e3-4aa6-b98c-50101f0229d9" (UID: "b096e2a5-08e3-4aa6-b98c-50101f0229d9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.163159 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b096e2a5-08e3-4aa6-b98c-50101f0229d9" (UID: "b096e2a5-08e3-4aa6-b98c-50101f0229d9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.164628 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b096e2a5-08e3-4aa6-b98c-50101f0229d9" (UID: "b096e2a5-08e3-4aa6-b98c-50101f0229d9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.181429 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.181454 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.181463 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.181472 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b096e2a5-08e3-4aa6-b98c-50101f0229d9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.298687 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-czbjh"] Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.301261 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" event={"ID":"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4","Type":"ContainerStarted","Data":"244dc2182531731fd4aadd9e00f887d0d25b4e5e95185568fecde620a28215e6"} Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.316680 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.316788 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-gk7lw" event={"ID":"b096e2a5-08e3-4aa6-b98c-50101f0229d9","Type":"ContainerDied","Data":"badd1da4f9256372804f64e41caaccedc52c06362a8e26b7a2ff2b6327ace719"} Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.316852 4726 scope.go:117] "RemoveContainer" containerID="2d5a1ac6ed14d039f09b5804239e2481024122e6b4f674d144d55d452b18e175" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.343393 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3c580861-1c2f-4f6b-933a-5203bd970f8a","Type":"ContainerStarted","Data":"bde78428d612a26397f0f8b26a36e721059024cbaebc999d9de1848149132f8a"} Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.353620 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e86b00b2-1546-4cfc-a51a-23b92e950bc8","Type":"ContainerStarted","Data":"26e26efac83a73711399717f3eebc7c95f8b3f47742d885efa2b4b5d5d3eb62f"} Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.372296 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc64455bb-r2zhn" event={"ID":"ebde1298-9550-4cac-a7a4-a4f31135b59f","Type":"ContainerStarted","Data":"221cb6163476026452ea09ec2f2660001a03106ce1256f0cc5030dcb246a57dc"} Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.373066 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.373091 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.429305 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-gk7lw"] Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.473372 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-gk7lw"] Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.482255 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7bc64455bb-r2zhn" podStartSLOduration=3.482238935 podStartE2EDuration="3.482238935s" podCreationTimestamp="2025-10-01 16:17:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:17:59.41556466 +0000 UTC m=+1132.317117237" watchObservedRunningTime="2025-10-01 16:17:59.482238935 +0000 UTC m=+1132.383791512" Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.569992 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-566c76dfb4-x49km"] Oct 01 16:17:59 crc kubenswrapper[4726]: I1001 16:17:59.823260 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b096e2a5-08e3-4aa6-b98c-50101f0229d9" path="/var/lib/kubelet/pods/b096e2a5-08e3-4aa6-b98c-50101f0229d9/volumes" Oct 01 16:18:00 crc kubenswrapper[4726]: I1001 16:18:00.369891 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:18:00 crc kubenswrapper[4726]: I1001 16:18:00.423777 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-566c76dfb4-x49km" event={"ID":"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984","Type":"ContainerStarted","Data":"de92528fc15954f4d81471dcab7ecafea712169a7ef79c4915d1d95a820c391e"} Oct 01 16:18:00 crc kubenswrapper[4726]: I1001 16:18:00.425778 4726 generic.go:334] "Generic (PLEG): container finished" podID="d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" containerID="671c329f9be6ecf907b98714e983c08724dc01956ed7e1321a43622a60724b1c" exitCode=0 Oct 01 16:18:00 crc kubenswrapper[4726]: I1001 16:18:00.425828 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-czbjh" event={"ID":"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e","Type":"ContainerDied","Data":"671c329f9be6ecf907b98714e983c08724dc01956ed7e1321a43622a60724b1c"} Oct 01 16:18:00 crc kubenswrapper[4726]: I1001 16:18:00.425845 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-czbjh" event={"ID":"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e","Type":"ContainerStarted","Data":"dd395c66044a73144b33efab265ad3bfc8f1c2fe11e20bdd847657ae2b3307ea"} Oct 01 16:18:00 crc kubenswrapper[4726]: I1001 16:18:00.443846 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3c580861-1c2f-4f6b-933a-5203bd970f8a","Type":"ContainerStarted","Data":"bceef75931d839c346375f01b154ac35c280a5726259675f7f0473f501361202"} Oct 01 16:18:00 crc kubenswrapper[4726]: I1001 16:18:00.447838 4726 generic.go:334] "Generic (PLEG): container finished" podID="d3f9ce65-7910-4afc-813a-5b1a9c88d6a4" containerID="8bf9ce7d3ebda7488d4ba054f7f3fd8f2cc5d1cb4e231135335f852382eb05be" exitCode=0 Oct 01 16:18:00 crc kubenswrapper[4726]: I1001 16:18:00.448520 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" event={"ID":"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4","Type":"ContainerDied","Data":"8bf9ce7d3ebda7488d4ba054f7f3fd8f2cc5d1cb4e231135335f852382eb05be"} Oct 01 16:18:00 crc kubenswrapper[4726]: I1001 16:18:00.643498 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:18:00 crc kubenswrapper[4726]: I1001 16:18:00.643619 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-68f479c67b-gj9j5" Oct 01 16:18:00 crc kubenswrapper[4726]: I1001 16:18:00.717937 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-75bf4bdcc5-mbcpq" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.472855 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.506322 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e86b00b2-1546-4cfc-a51a-23b92e950bc8","Type":"ContainerStarted","Data":"40826bad473aa6e993f5c08de182c0148854cccd39f29b519a8823126524c6ea"} Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.511642 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" event={"ID":"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4","Type":"ContainerDied","Data":"244dc2182531731fd4aadd9e00f887d0d25b4e5e95185568fecde620a28215e6"} Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.511685 4726 scope.go:117] "RemoveContainer" containerID="8bf9ce7d3ebda7488d4ba054f7f3fd8f2cc5d1cb4e231135335f852382eb05be" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.511771 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.564083 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-ovsdbserver-sb\") pod \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.564151 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-dns-swift-storage-0\") pod \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.564264 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-ovsdbserver-nb\") pod \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.564306 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-dns-svc\") pod \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.564325 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzmw9\" (UniqueName: \"kubernetes.io/projected/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-kube-api-access-gzmw9\") pod \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.564428 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-config\") pod \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\" (UID: \"d3f9ce65-7910-4afc-813a-5b1a9c88d6a4\") " Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.573505 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-kube-api-access-gzmw9" (OuterVolumeSpecName: "kube-api-access-gzmw9") pod "d3f9ce65-7910-4afc-813a-5b1a9c88d6a4" (UID: "d3f9ce65-7910-4afc-813a-5b1a9c88d6a4"). InnerVolumeSpecName "kube-api-access-gzmw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.586298 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d3f9ce65-7910-4afc-813a-5b1a9c88d6a4" (UID: "d3f9ce65-7910-4afc-813a-5b1a9c88d6a4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.597681 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d3f9ce65-7910-4afc-813a-5b1a9c88d6a4" (UID: "d3f9ce65-7910-4afc-813a-5b1a9c88d6a4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.605540 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d3f9ce65-7910-4afc-813a-5b1a9c88d6a4" (UID: "d3f9ce65-7910-4afc-813a-5b1a9c88d6a4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.612203 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d3f9ce65-7910-4afc-813a-5b1a9c88d6a4" (UID: "d3f9ce65-7910-4afc-813a-5b1a9c88d6a4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.619324 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-config" (OuterVolumeSpecName: "config") pod "d3f9ce65-7910-4afc-813a-5b1a9c88d6a4" (UID: "d3f9ce65-7910-4afc-813a-5b1a9c88d6a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.666211 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.666241 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.666253 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.666263 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.666272 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.666281 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzmw9\" (UniqueName: \"kubernetes.io/projected/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4-kube-api-access-gzmw9\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.899421 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf"] Oct 01 16:18:01 crc kubenswrapper[4726]: I1001 16:18:01.900173 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cc8b5d5c5-lw9xf"] Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.522803 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-99545b666-4wmpl" event={"ID":"7f25ba81-3f98-4d63-8035-2b30a1df5f54","Type":"ContainerStarted","Data":"17e1c293e97dc4b634a74a00a4c0af7fcfe7c2d885af38e293fe750ad3612f19"} Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.523100 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-99545b666-4wmpl" event={"ID":"7f25ba81-3f98-4d63-8035-2b30a1df5f54","Type":"ContainerStarted","Data":"88ea0c65747b15a817175e7a0374f0240010fd241fd09112772bd24fca203ece"} Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.525408 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-566c76dfb4-x49km" event={"ID":"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984","Type":"ContainerStarted","Data":"eae9e5f7a8a4413ed91ee774a7f3ec41fe23fab3f71b4aca60e097647f504fb5"} Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.525464 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-566c76dfb4-x49km" event={"ID":"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984","Type":"ContainerStarted","Data":"bd05040824b5af9e2be2be07f72261af0136cfbbfca823ddfc3960a35db693bf"} Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.525552 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.527085 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-czbjh" event={"ID":"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e","Type":"ContainerStarted","Data":"b3856e2da018ee9d1b6a1b9c847e0df283e574a77b4f04d7320b96e91a935ce7"} Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.527205 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.530347 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3c580861-1c2f-4f6b-933a-5203bd970f8a","Type":"ContainerStarted","Data":"f32d21061674e2f35dc702c02fa944d6294b6dc7f17331bf978e86d289571bdd"} Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.530500 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3c580861-1c2f-4f6b-933a-5203bd970f8a" containerName="cinder-api-log" containerID="cri-o://bceef75931d839c346375f01b154ac35c280a5726259675f7f0473f501361202" gracePeriod=30 Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.530687 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.530685 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3c580861-1c2f-4f6b-933a-5203bd970f8a" containerName="cinder-api" containerID="cri-o://f32d21061674e2f35dc702c02fa944d6294b6dc7f17331bf978e86d289571bdd" gracePeriod=30 Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.535980 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57cbb4948c-gcc56" event={"ID":"94005d17-9bb0-4039-b493-5057a554f054","Type":"ContainerStarted","Data":"4ed5af137ac145aafeafb37d4c51a08d0081738fbb7cda6d21eaec94a11bc6c4"} Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.536021 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57cbb4948c-gcc56" event={"ID":"94005d17-9bb0-4039-b493-5057a554f054","Type":"ContainerStarted","Data":"ee7a322df8fa4e4bfc43fe95093575d1c399b8284f07bb34405471297372067b"} Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.539731 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e86b00b2-1546-4cfc-a51a-23b92e950bc8","Type":"ContainerStarted","Data":"857e05b3269cf36b182c4e6368942373520c3bc3f6759e90b5b9e98184750b1d"} Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.548128 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-99545b666-4wmpl" podStartSLOduration=2.802388813 podStartE2EDuration="6.548107294s" podCreationTimestamp="2025-10-01 16:17:56 +0000 UTC" firstStartedPulling="2025-10-01 16:17:57.60189357 +0000 UTC m=+1130.503446147" lastFinishedPulling="2025-10-01 16:18:01.347612051 +0000 UTC m=+1134.249164628" observedRunningTime="2025-10-01 16:18:02.545977414 +0000 UTC m=+1135.447529991" watchObservedRunningTime="2025-10-01 16:18:02.548107294 +0000 UTC m=+1135.449659871" Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.578378 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-czbjh" podStartSLOduration=4.578361179 podStartE2EDuration="4.578361179s" podCreationTimestamp="2025-10-01 16:17:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:18:02.576131405 +0000 UTC m=+1135.477683982" watchObservedRunningTime="2025-10-01 16:18:02.578361179 +0000 UTC m=+1135.479913756" Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.621564 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-57cbb4948c-gcc56" podStartSLOduration=2.717819557 podStartE2EDuration="6.621548472s" podCreationTimestamp="2025-10-01 16:17:56 +0000 UTC" firstStartedPulling="2025-10-01 16:17:57.439263654 +0000 UTC m=+1130.340816231" lastFinishedPulling="2025-10-01 16:18:01.342992569 +0000 UTC m=+1134.244545146" observedRunningTime="2025-10-01 16:18:02.604085283 +0000 UTC m=+1135.505637870" watchObservedRunningTime="2025-10-01 16:18:02.621548472 +0000 UTC m=+1135.523101039" Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.626397 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.853184983 podStartE2EDuration="5.62638504s" podCreationTimestamp="2025-10-01 16:17:57 +0000 UTC" firstStartedPulling="2025-10-01 16:17:58.669013874 +0000 UTC m=+1131.570566451" lastFinishedPulling="2025-10-01 16:17:59.442213941 +0000 UTC m=+1132.343766508" observedRunningTime="2025-10-01 16:18:02.619808263 +0000 UTC m=+1135.521360840" watchObservedRunningTime="2025-10-01 16:18:02.62638504 +0000 UTC m=+1135.527937617" Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.666102 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.666076654 podStartE2EDuration="5.666076654s" podCreationTimestamp="2025-10-01 16:17:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:18:02.643295364 +0000 UTC m=+1135.544847941" watchObservedRunningTime="2025-10-01 16:18:02.666076654 +0000 UTC m=+1135.567629231" Oct 01 16:18:02 crc kubenswrapper[4726]: I1001 16:18:02.685325 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-566c76dfb4-x49km" podStartSLOduration=4.685307284 podStartE2EDuration="4.685307284s" podCreationTimestamp="2025-10-01 16:17:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:18:02.682542715 +0000 UTC m=+1135.584095292" watchObservedRunningTime="2025-10-01 16:18:02.685307284 +0000 UTC m=+1135.586859861" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.074437 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.561981 4726 generic.go:334] "Generic (PLEG): container finished" podID="3c580861-1c2f-4f6b-933a-5203bd970f8a" containerID="f32d21061674e2f35dc702c02fa944d6294b6dc7f17331bf978e86d289571bdd" exitCode=0 Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.562016 4726 generic.go:334] "Generic (PLEG): container finished" podID="3c580861-1c2f-4f6b-933a-5203bd970f8a" containerID="bceef75931d839c346375f01b154ac35c280a5726259675f7f0473f501361202" exitCode=143 Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.562164 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3c580861-1c2f-4f6b-933a-5203bd970f8a","Type":"ContainerDied","Data":"f32d21061674e2f35dc702c02fa944d6294b6dc7f17331bf978e86d289571bdd"} Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.562210 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3c580861-1c2f-4f6b-933a-5203bd970f8a","Type":"ContainerDied","Data":"bceef75931d839c346375f01b154ac35c280a5726259675f7f0473f501361202"} Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.755933 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 01 16:18:03 crc kubenswrapper[4726]: E1001 16:18:03.756804 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f9ce65-7910-4afc-813a-5b1a9c88d6a4" containerName="init" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.756826 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f9ce65-7910-4afc-813a-5b1a9c88d6a4" containerName="init" Oct 01 16:18:03 crc kubenswrapper[4726]: E1001 16:18:03.756879 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b096e2a5-08e3-4aa6-b98c-50101f0229d9" containerName="init" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.756888 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b096e2a5-08e3-4aa6-b98c-50101f0229d9" containerName="init" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.757132 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f9ce65-7910-4afc-813a-5b1a9c88d6a4" containerName="init" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.757153 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b096e2a5-08e3-4aa6-b98c-50101f0229d9" containerName="init" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.757912 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.763275 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-fjnmt" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.763469 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.763599 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.769776 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.873754 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3f9ce65-7910-4afc-813a-5b1a9c88d6a4" path="/var/lib/kubelet/pods/d3f9ce65-7910-4afc-813a-5b1a9c88d6a4/volumes" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.931951 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/37c371f5-1756-49e1-82ec-40b7cb86f524-openstack-config-secret\") pod \"openstackclient\" (UID: \"37c371f5-1756-49e1-82ec-40b7cb86f524\") " pod="openstack/openstackclient" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.932081 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37c371f5-1756-49e1-82ec-40b7cb86f524-combined-ca-bundle\") pod \"openstackclient\" (UID: \"37c371f5-1756-49e1-82ec-40b7cb86f524\") " pod="openstack/openstackclient" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.932176 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/37c371f5-1756-49e1-82ec-40b7cb86f524-openstack-config\") pod \"openstackclient\" (UID: \"37c371f5-1756-49e1-82ec-40b7cb86f524\") " pod="openstack/openstackclient" Oct 01 16:18:03 crc kubenswrapper[4726]: I1001 16:18:03.932362 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r68vr\" (UniqueName: \"kubernetes.io/projected/37c371f5-1756-49e1-82ec-40b7cb86f524-kube-api-access-r68vr\") pod \"openstackclient\" (UID: \"37c371f5-1756-49e1-82ec-40b7cb86f524\") " pod="openstack/openstackclient" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.034362 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r68vr\" (UniqueName: \"kubernetes.io/projected/37c371f5-1756-49e1-82ec-40b7cb86f524-kube-api-access-r68vr\") pod \"openstackclient\" (UID: \"37c371f5-1756-49e1-82ec-40b7cb86f524\") " pod="openstack/openstackclient" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.034456 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/37c371f5-1756-49e1-82ec-40b7cb86f524-openstack-config-secret\") pod \"openstackclient\" (UID: \"37c371f5-1756-49e1-82ec-40b7cb86f524\") " pod="openstack/openstackclient" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.034507 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37c371f5-1756-49e1-82ec-40b7cb86f524-combined-ca-bundle\") pod \"openstackclient\" (UID: \"37c371f5-1756-49e1-82ec-40b7cb86f524\") " pod="openstack/openstackclient" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.034568 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/37c371f5-1756-49e1-82ec-40b7cb86f524-openstack-config\") pod \"openstackclient\" (UID: \"37c371f5-1756-49e1-82ec-40b7cb86f524\") " pod="openstack/openstackclient" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.035640 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/37c371f5-1756-49e1-82ec-40b7cb86f524-openstack-config\") pod \"openstackclient\" (UID: \"37c371f5-1756-49e1-82ec-40b7cb86f524\") " pod="openstack/openstackclient" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.043564 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/37c371f5-1756-49e1-82ec-40b7cb86f524-openstack-config-secret\") pod \"openstackclient\" (UID: \"37c371f5-1756-49e1-82ec-40b7cb86f524\") " pod="openstack/openstackclient" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.055298 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37c371f5-1756-49e1-82ec-40b7cb86f524-combined-ca-bundle\") pod \"openstackclient\" (UID: \"37c371f5-1756-49e1-82ec-40b7cb86f524\") " pod="openstack/openstackclient" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.056297 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r68vr\" (UniqueName: \"kubernetes.io/projected/37c371f5-1756-49e1-82ec-40b7cb86f524-kube-api-access-r68vr\") pod \"openstackclient\" (UID: \"37c371f5-1756-49e1-82ec-40b7cb86f524\") " pod="openstack/openstackclient" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.079803 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.183996 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.287139 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-598fc759b9-28cln"] Oct 01 16:18:04 crc kubenswrapper[4726]: E1001 16:18:04.287790 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c580861-1c2f-4f6b-933a-5203bd970f8a" containerName="cinder-api-log" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.287809 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c580861-1c2f-4f6b-933a-5203bd970f8a" containerName="cinder-api-log" Oct 01 16:18:04 crc kubenswrapper[4726]: E1001 16:18:04.287823 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c580861-1c2f-4f6b-933a-5203bd970f8a" containerName="cinder-api" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.287830 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c580861-1c2f-4f6b-933a-5203bd970f8a" containerName="cinder-api" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.288010 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c580861-1c2f-4f6b-933a-5203bd970f8a" containerName="cinder-api" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.288033 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c580861-1c2f-4f6b-933a-5203bd970f8a" containerName="cinder-api-log" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.288992 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.299401 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.299540 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.302344 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-598fc759b9-28cln"] Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.342816 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-scripts\") pod \"3c580861-1c2f-4f6b-933a-5203bd970f8a\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.342895 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-combined-ca-bundle\") pod \"3c580861-1c2f-4f6b-933a-5203bd970f8a\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.342938 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-config-data\") pod \"3c580861-1c2f-4f6b-933a-5203bd970f8a\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.342966 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c580861-1c2f-4f6b-933a-5203bd970f8a-etc-machine-id\") pod \"3c580861-1c2f-4f6b-933a-5203bd970f8a\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.343012 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p7lx\" (UniqueName: \"kubernetes.io/projected/3c580861-1c2f-4f6b-933a-5203bd970f8a-kube-api-access-4p7lx\") pod \"3c580861-1c2f-4f6b-933a-5203bd970f8a\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.343040 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-config-data-custom\") pod \"3c580861-1c2f-4f6b-933a-5203bd970f8a\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.343139 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c580861-1c2f-4f6b-933a-5203bd970f8a-logs\") pod \"3c580861-1c2f-4f6b-933a-5203bd970f8a\" (UID: \"3c580861-1c2f-4f6b-933a-5203bd970f8a\") " Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.345223 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c580861-1c2f-4f6b-933a-5203bd970f8a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3c580861-1c2f-4f6b-933a-5203bd970f8a" (UID: "3c580861-1c2f-4f6b-933a-5203bd970f8a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.349362 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c580861-1c2f-4f6b-933a-5203bd970f8a-kube-api-access-4p7lx" (OuterVolumeSpecName: "kube-api-access-4p7lx") pod "3c580861-1c2f-4f6b-933a-5203bd970f8a" (UID: "3c580861-1c2f-4f6b-933a-5203bd970f8a"). InnerVolumeSpecName "kube-api-access-4p7lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.362882 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-scripts" (OuterVolumeSpecName: "scripts") pod "3c580861-1c2f-4f6b-933a-5203bd970f8a" (UID: "3c580861-1c2f-4f6b-933a-5203bd970f8a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.363191 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c580861-1c2f-4f6b-933a-5203bd970f8a-logs" (OuterVolumeSpecName: "logs") pod "3c580861-1c2f-4f6b-933a-5203bd970f8a" (UID: "3c580861-1c2f-4f6b-933a-5203bd970f8a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.368723 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3c580861-1c2f-4f6b-933a-5203bd970f8a" (UID: "3c580861-1c2f-4f6b-933a-5203bd970f8a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.412965 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c580861-1c2f-4f6b-933a-5203bd970f8a" (UID: "3c580861-1c2f-4f6b-933a-5203bd970f8a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.431152 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-config-data" (OuterVolumeSpecName: "config-data") pod "3c580861-1c2f-4f6b-933a-5203bd970f8a" (UID: "3c580861-1c2f-4f6b-933a-5203bd970f8a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.446850 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-config\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.446932 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-httpd-config\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.447007 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-internal-tls-certs\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.447358 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-combined-ca-bundle\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.447427 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-public-tls-certs\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.447513 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-ovndb-tls-certs\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.447598 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqgpq\" (UniqueName: \"kubernetes.io/projected/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-kube-api-access-vqgpq\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.447716 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p7lx\" (UniqueName: \"kubernetes.io/projected/3c580861-1c2f-4f6b-933a-5203bd970f8a-kube-api-access-4p7lx\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.447740 4726 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.447758 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c580861-1c2f-4f6b-933a-5203bd970f8a-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.447775 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.447792 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.447811 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c580861-1c2f-4f6b-933a-5203bd970f8a-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.447827 4726 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c580861-1c2f-4f6b-933a-5203bd970f8a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.549194 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-ovndb-tls-certs\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.549269 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqgpq\" (UniqueName: \"kubernetes.io/projected/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-kube-api-access-vqgpq\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.549319 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-config\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.549342 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-httpd-config\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.549373 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-internal-tls-certs\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.549406 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-combined-ca-bundle\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.549430 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-public-tls-certs\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.554943 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-ovndb-tls-certs\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.556350 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-internal-tls-certs\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.556498 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-public-tls-certs\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.557837 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-config\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.558101 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-httpd-config\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.559666 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-combined-ca-bundle\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.571262 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.574392 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqgpq\" (UniqueName: \"kubernetes.io/projected/16ed9afc-f760-4a59-8db3-8d76c1bd42fa-kube-api-access-vqgpq\") pod \"neutron-598fc759b9-28cln\" (UID: \"16ed9afc-f760-4a59-8db3-8d76c1bd42fa\") " pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.574448 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3c580861-1c2f-4f6b-933a-5203bd970f8a","Type":"ContainerDied","Data":"bde78428d612a26397f0f8b26a36e721059024cbaebc999d9de1848149132f8a"} Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.574488 4726 scope.go:117] "RemoveContainer" containerID="f32d21061674e2f35dc702c02fa944d6294b6dc7f17331bf978e86d289571bdd" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.614308 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.624264 4726 scope.go:117] "RemoveContainer" containerID="bceef75931d839c346375f01b154ac35c280a5726259675f7f0473f501361202" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.628333 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.647156 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.653582 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.664026 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.664324 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.667474 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.674293 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.674589 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.674809 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.753641 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.753739 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.753776 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.753806 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-config-data-custom\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.753827 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-scripts\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.753855 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-logs\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.753873 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-config-data\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.753907 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.753946 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk28k\" (UniqueName: \"kubernetes.io/projected/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-kube-api-access-xk28k\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.856103 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.856184 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.856246 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-config-data-custom\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.856282 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-scripts\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.856333 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-logs\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.856340 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.856367 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-config-data\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.856469 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.856541 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk28k\" (UniqueName: \"kubernetes.io/projected/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-kube-api-access-xk28k\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.856677 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.857708 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-logs\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.863149 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.864102 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-scripts\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.867665 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.868006 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-config-data-custom\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.870887 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-config-data\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.871343 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:04 crc kubenswrapper[4726]: I1001 16:18:04.878288 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk28k\" (UniqueName: \"kubernetes.io/projected/20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7-kube-api-access-xk28k\") pod \"cinder-api-0\" (UID: \"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7\") " pod="openstack/cinder-api-0" Oct 01 16:18:05 crc kubenswrapper[4726]: I1001 16:18:05.005191 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 16:18:05 crc kubenswrapper[4726]: I1001 16:18:05.320860 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-598fc759b9-28cln"] Oct 01 16:18:05 crc kubenswrapper[4726]: I1001 16:18:05.631914 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-598fc759b9-28cln" event={"ID":"16ed9afc-f760-4a59-8db3-8d76c1bd42fa","Type":"ContainerStarted","Data":"6187d4269e8f4b19d083d345ea33c4305c0debbe8ef602e20cb2d70c84f9a309"} Oct 01 16:18:05 crc kubenswrapper[4726]: I1001 16:18:05.634805 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:18:05 crc kubenswrapper[4726]: I1001 16:18:05.657142 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"37c371f5-1756-49e1-82ec-40b7cb86f524","Type":"ContainerStarted","Data":"7d00c722ef0044e261c2f1634a4034559f9397fd29c0be43b9daebada121b396"} Oct 01 16:18:05 crc kubenswrapper[4726]: I1001 16:18:05.817304 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c580861-1c2f-4f6b-933a-5203bd970f8a" path="/var/lib/kubelet/pods/3c580861-1c2f-4f6b-933a-5203bd970f8a/volumes" Oct 01 16:18:06 crc kubenswrapper[4726]: I1001 16:18:06.672861 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-598fc759b9-28cln" event={"ID":"16ed9afc-f760-4a59-8db3-8d76c1bd42fa","Type":"ContainerStarted","Data":"35a112913a60182e2d8377f699577fe5328b459f4810ac3238f2cb24998ca404"} Oct 01 16:18:06 crc kubenswrapper[4726]: I1001 16:18:06.673405 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:06 crc kubenswrapper[4726]: I1001 16:18:06.673419 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-598fc759b9-28cln" event={"ID":"16ed9afc-f760-4a59-8db3-8d76c1bd42fa","Type":"ContainerStarted","Data":"d1d03b8b4ab07d562c047475bd07256fc2bd1599f99d879d6c80b7cee1490a6c"} Oct 01 16:18:06 crc kubenswrapper[4726]: I1001 16:18:06.678066 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7","Type":"ContainerStarted","Data":"afc18c6e525b55097e23acaae6639f777238923b53d6ed7f3aca5f70d0b37566"} Oct 01 16:18:06 crc kubenswrapper[4726]: I1001 16:18:06.678111 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7","Type":"ContainerStarted","Data":"56ba82745f03ffa93bad31e468c09eb4e9e5999da59c2df2c7777fc17c6801a7"} Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.370532 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-598fc759b9-28cln" podStartSLOduration=3.370511442 podStartE2EDuration="3.370511442s" podCreationTimestamp="2025-10-01 16:18:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:18:06.704346292 +0000 UTC m=+1139.605898869" watchObservedRunningTime="2025-10-01 16:18:07.370511442 +0000 UTC m=+1140.272064009" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.376561 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5649c4d89d-cjbv9"] Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.378632 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.381274 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.381430 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.400970 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5649c4d89d-cjbv9"] Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.518848 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-config-data-custom\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.518902 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t59f\" (UniqueName: \"kubernetes.io/projected/e28be6af-1011-49d4-a949-77fdf75e07d9-kube-api-access-7t59f\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.518928 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-combined-ca-bundle\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.519087 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e28be6af-1011-49d4-a949-77fdf75e07d9-logs\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.519246 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-config-data\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.519290 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-public-tls-certs\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.519313 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-internal-tls-certs\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.620704 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t59f\" (UniqueName: \"kubernetes.io/projected/e28be6af-1011-49d4-a949-77fdf75e07d9-kube-api-access-7t59f\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.620766 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-combined-ca-bundle\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.620802 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e28be6af-1011-49d4-a949-77fdf75e07d9-logs\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.620858 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-config-data\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.620894 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-public-tls-certs\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.620921 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-internal-tls-certs\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.621043 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-config-data-custom\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.621331 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e28be6af-1011-49d4-a949-77fdf75e07d9-logs\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.627335 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-config-data-custom\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.628711 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-combined-ca-bundle\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.633378 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-config-data\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.645182 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-internal-tls-certs\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.650021 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28be6af-1011-49d4-a949-77fdf75e07d9-public-tls-certs\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.654546 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t59f\" (UniqueName: \"kubernetes.io/projected/e28be6af-1011-49d4-a949-77fdf75e07d9-kube-api-access-7t59f\") pod \"barbican-api-5649c4d89d-cjbv9\" (UID: \"e28be6af-1011-49d4-a949-77fdf75e07d9\") " pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.697956 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.738975 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7","Type":"ContainerStarted","Data":"d0e9725853cb8d1be27092c215e9358aa94b4cb4e4576321f5275c2e9e825c8e"} Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.739069 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 01 16:18:07 crc kubenswrapper[4726]: I1001 16:18:07.789796 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.789775958 podStartE2EDuration="3.789775958s" podCreationTimestamp="2025-10-01 16:18:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:18:07.786514934 +0000 UTC m=+1140.688067501" watchObservedRunningTime="2025-10-01 16:18:07.789775958 +0000 UTC m=+1140.691328535" Oct 01 16:18:08 crc kubenswrapper[4726]: I1001 16:18:08.385286 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5649c4d89d-cjbv9"] Oct 01 16:18:08 crc kubenswrapper[4726]: I1001 16:18:08.540758 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 01 16:18:08 crc kubenswrapper[4726]: I1001 16:18:08.587021 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:18:08 crc kubenswrapper[4726]: I1001 16:18:08.797395 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e86b00b2-1546-4cfc-a51a-23b92e950bc8" containerName="cinder-scheduler" containerID="cri-o://40826bad473aa6e993f5c08de182c0148854cccd39f29b519a8823126524c6ea" gracePeriod=30 Oct 01 16:18:08 crc kubenswrapper[4726]: I1001 16:18:08.797704 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5649c4d89d-cjbv9" event={"ID":"e28be6af-1011-49d4-a949-77fdf75e07d9","Type":"ContainerStarted","Data":"315412939c22e5067664275ee7d383d116b4dc02f074e5a8b021c9f0b6b3ada0"} Oct 01 16:18:08 crc kubenswrapper[4726]: I1001 16:18:08.797727 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5649c4d89d-cjbv9" event={"ID":"e28be6af-1011-49d4-a949-77fdf75e07d9","Type":"ContainerStarted","Data":"0ab1b1ef6e727673a5c8d5059f2cc9c28efc52baf3a1b8b034bc9ea7216e4099"} Oct 01 16:18:08 crc kubenswrapper[4726]: I1001 16:18:08.798847 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e86b00b2-1546-4cfc-a51a-23b92e950bc8" containerName="probe" containerID="cri-o://857e05b3269cf36b182c4e6368942373520c3bc3f6759e90b5b9e98184750b1d" gracePeriod=30 Oct 01 16:18:08 crc kubenswrapper[4726]: I1001 16:18:08.951179 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:18:08 crc kubenswrapper[4726]: I1001 16:18:08.997976 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bdn6t"] Oct 01 16:18:08 crc kubenswrapper[4726]: I1001 16:18:08.998210 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" podUID="34e6cb54-5cda-4ac4-bfe1-e59986daa349" containerName="dnsmasq-dns" containerID="cri-o://c82e46e2d4b5502528f154f2d29fabd0fe75b2be4ccdfcaee912aad6362e7885" gracePeriod=10 Oct 01 16:18:09 crc kubenswrapper[4726]: I1001 16:18:09.544851 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:18:09 crc kubenswrapper[4726]: I1001 16:18:09.819200 4726 generic.go:334] "Generic (PLEG): container finished" podID="34e6cb54-5cda-4ac4-bfe1-e59986daa349" containerID="c82e46e2d4b5502528f154f2d29fabd0fe75b2be4ccdfcaee912aad6362e7885" exitCode=0 Oct 01 16:18:09 crc kubenswrapper[4726]: I1001 16:18:09.830326 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" event={"ID":"34e6cb54-5cda-4ac4-bfe1-e59986daa349","Type":"ContainerDied","Data":"c82e46e2d4b5502528f154f2d29fabd0fe75b2be4ccdfcaee912aad6362e7885"} Oct 01 16:18:09 crc kubenswrapper[4726]: I1001 16:18:09.831974 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5649c4d89d-cjbv9" event={"ID":"e28be6af-1011-49d4-a949-77fdf75e07d9","Type":"ContainerStarted","Data":"c1223ded95ff67f0937c97a39f2e1e165361cce44d996fdca1b9350b0b41c7d3"} Oct 01 16:18:09 crc kubenswrapper[4726]: I1001 16:18:09.832094 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:09 crc kubenswrapper[4726]: I1001 16:18:09.832125 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:09 crc kubenswrapper[4726]: I1001 16:18:09.841140 4726 generic.go:334] "Generic (PLEG): container finished" podID="e86b00b2-1546-4cfc-a51a-23b92e950bc8" containerID="857e05b3269cf36b182c4e6368942373520c3bc3f6759e90b5b9e98184750b1d" exitCode=0 Oct 01 16:18:09 crc kubenswrapper[4726]: I1001 16:18:09.841219 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e86b00b2-1546-4cfc-a51a-23b92e950bc8","Type":"ContainerDied","Data":"857e05b3269cf36b182c4e6368942373520c3bc3f6759e90b5b9e98184750b1d"} Oct 01 16:18:09 crc kubenswrapper[4726]: I1001 16:18:09.858461 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5649c4d89d-cjbv9" podStartSLOduration=2.858344259 podStartE2EDuration="2.858344259s" podCreationTimestamp="2025-10-01 16:18:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:18:09.852662397 +0000 UTC m=+1142.754214974" watchObservedRunningTime="2025-10-01 16:18:09.858344259 +0000 UTC m=+1142.759904846" Oct 01 16:18:09 crc kubenswrapper[4726]: I1001 16:18:09.973550 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.095067 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.188081 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-ovsdbserver-nb\") pod \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.188233 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-ovsdbserver-sb\") pod \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.188272 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-dns-swift-storage-0\") pod \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.188311 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-dns-svc\") pod \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.188434 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zqpd\" (UniqueName: \"kubernetes.io/projected/34e6cb54-5cda-4ac4-bfe1-e59986daa349-kube-api-access-6zqpd\") pod \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.188459 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-config\") pod \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\" (UID: \"34e6cb54-5cda-4ac4-bfe1-e59986daa349\") " Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.210231 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34e6cb54-5cda-4ac4-bfe1-e59986daa349-kube-api-access-6zqpd" (OuterVolumeSpecName: "kube-api-access-6zqpd") pod "34e6cb54-5cda-4ac4-bfe1-e59986daa349" (UID: "34e6cb54-5cda-4ac4-bfe1-e59986daa349"). InnerVolumeSpecName "kube-api-access-6zqpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.245770 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "34e6cb54-5cda-4ac4-bfe1-e59986daa349" (UID: "34e6cb54-5cda-4ac4-bfe1-e59986daa349"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.252012 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-config" (OuterVolumeSpecName: "config") pod "34e6cb54-5cda-4ac4-bfe1-e59986daa349" (UID: "34e6cb54-5cda-4ac4-bfe1-e59986daa349"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.262454 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "34e6cb54-5cda-4ac4-bfe1-e59986daa349" (UID: "34e6cb54-5cda-4ac4-bfe1-e59986daa349"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.262607 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "34e6cb54-5cda-4ac4-bfe1-e59986daa349" (UID: "34e6cb54-5cda-4ac4-bfe1-e59986daa349"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.290482 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zqpd\" (UniqueName: \"kubernetes.io/projected/34e6cb54-5cda-4ac4-bfe1-e59986daa349-kube-api-access-6zqpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.290523 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.290536 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.290546 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.290583 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.313431 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "34e6cb54-5cda-4ac4-bfe1-e59986daa349" (UID: "34e6cb54-5cda-4ac4-bfe1-e59986daa349"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.392517 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34e6cb54-5cda-4ac4-bfe1-e59986daa349-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.851881 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" event={"ID":"34e6cb54-5cda-4ac4-bfe1-e59986daa349","Type":"ContainerDied","Data":"bbb9783fb8558f7a4742d4e955d0800fcdeec69656503992055bbba90628b288"} Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.851936 4726 scope.go:117] "RemoveContainer" containerID="c82e46e2d4b5502528f154f2d29fabd0fe75b2be4ccdfcaee912aad6362e7885" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.852103 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-bdn6t" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.878543 4726 scope.go:117] "RemoveContainer" containerID="833925b048845f5fafb64f4ad02060e065c312c00e235d9df20ec90a80631351" Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.886366 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bdn6t"] Oct 01 16:18:10 crc kubenswrapper[4726]: I1001 16:18:10.896067 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-bdn6t"] Oct 01 16:18:11 crc kubenswrapper[4726]: I1001 16:18:11.819921 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34e6cb54-5cda-4ac4-bfe1-e59986daa349" path="/var/lib/kubelet/pods/34e6cb54-5cda-4ac4-bfe1-e59986daa349/volumes" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.266870 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-57c9dff847-vbzrr"] Oct 01 16:18:12 crc kubenswrapper[4726]: E1001 16:18:12.267560 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34e6cb54-5cda-4ac4-bfe1-e59986daa349" containerName="init" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.267579 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="34e6cb54-5cda-4ac4-bfe1-e59986daa349" containerName="init" Oct 01 16:18:12 crc kubenswrapper[4726]: E1001 16:18:12.267603 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34e6cb54-5cda-4ac4-bfe1-e59986daa349" containerName="dnsmasq-dns" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.267610 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="34e6cb54-5cda-4ac4-bfe1-e59986daa349" containerName="dnsmasq-dns" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.267776 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="34e6cb54-5cda-4ac4-bfe1-e59986daa349" containerName="dnsmasq-dns" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.268751 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.274082 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.274231 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.274303 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.316190 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-57c9dff847-vbzrr"] Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.329096 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ac310a-31b1-40c9-9724-2c488c0061e2-log-httpd\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.329156 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ac310a-31b1-40c9-9724-2c488c0061e2-run-httpd\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.329182 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ac310a-31b1-40c9-9724-2c488c0061e2-public-tls-certs\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.329206 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ac310a-31b1-40c9-9724-2c488c0061e2-internal-tls-certs\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.329254 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85ac310a-31b1-40c9-9724-2c488c0061e2-config-data\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.329288 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ac310a-31b1-40c9-9724-2c488c0061e2-combined-ca-bundle\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.329336 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/85ac310a-31b1-40c9-9724-2c488c0061e2-etc-swift\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.329362 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd4zl\" (UniqueName: \"kubernetes.io/projected/85ac310a-31b1-40c9-9724-2c488c0061e2-kube-api-access-bd4zl\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.431403 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ac310a-31b1-40c9-9724-2c488c0061e2-run-httpd\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.431451 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ac310a-31b1-40c9-9724-2c488c0061e2-public-tls-certs\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.431473 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ac310a-31b1-40c9-9724-2c488c0061e2-internal-tls-certs\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.431523 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85ac310a-31b1-40c9-9724-2c488c0061e2-config-data\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.431559 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ac310a-31b1-40c9-9724-2c488c0061e2-combined-ca-bundle\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.431586 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/85ac310a-31b1-40c9-9724-2c488c0061e2-etc-swift\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.431619 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd4zl\" (UniqueName: \"kubernetes.io/projected/85ac310a-31b1-40c9-9724-2c488c0061e2-kube-api-access-bd4zl\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.431661 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ac310a-31b1-40c9-9724-2c488c0061e2-log-httpd\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.431936 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ac310a-31b1-40c9-9724-2c488c0061e2-run-httpd\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.432018 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ac310a-31b1-40c9-9724-2c488c0061e2-log-httpd\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.437855 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ac310a-31b1-40c9-9724-2c488c0061e2-public-tls-certs\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.438279 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85ac310a-31b1-40c9-9724-2c488c0061e2-config-data\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.438734 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85ac310a-31b1-40c9-9724-2c488c0061e2-internal-tls-certs\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.442892 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/85ac310a-31b1-40c9-9724-2c488c0061e2-etc-swift\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.448105 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ac310a-31b1-40c9-9724-2c488c0061e2-combined-ca-bundle\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.463730 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd4zl\" (UniqueName: \"kubernetes.io/projected/85ac310a-31b1-40c9-9724-2c488c0061e2-kube-api-access-bd4zl\") pod \"swift-proxy-57c9dff847-vbzrr\" (UID: \"85ac310a-31b1-40c9-9724-2c488c0061e2\") " pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:12 crc kubenswrapper[4726]: I1001 16:18:12.589642 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:13 crc kubenswrapper[4726]: I1001 16:18:13.492956 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:13 crc kubenswrapper[4726]: I1001 16:18:13.493502 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="ceilometer-central-agent" containerID="cri-o://5c2a0c28fb1998f178aec0b143da14133c36382949f778b4035b4bb686b5d029" gracePeriod=30 Oct 01 16:18:13 crc kubenswrapper[4726]: I1001 16:18:13.494066 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="ceilometer-notification-agent" containerID="cri-o://4afa9e937c19dec8329428804da5e638a30d95fb07f42650d2de6a12ddec0eab" gracePeriod=30 Oct 01 16:18:13 crc kubenswrapper[4726]: I1001 16:18:13.494076 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="sg-core" containerID="cri-o://73c80c2c9d0d8cfaa3cca243004a745910379c837d85a4e6bce907ed5e68bf1c" gracePeriod=30 Oct 01 16:18:13 crc kubenswrapper[4726]: I1001 16:18:13.497306 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="proxy-httpd" containerID="cri-o://afd47f4eca931d1883629f439f47d59026063730435b0d205c5e545731a91049" gracePeriod=30 Oct 01 16:18:13 crc kubenswrapper[4726]: I1001 16:18:13.504006 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.154:3000/\": EOF" Oct 01 16:18:13 crc kubenswrapper[4726]: I1001 16:18:13.888404 4726 generic.go:334] "Generic (PLEG): container finished" podID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerID="afd47f4eca931d1883629f439f47d59026063730435b0d205c5e545731a91049" exitCode=0 Oct 01 16:18:13 crc kubenswrapper[4726]: I1001 16:18:13.888657 4726 generic.go:334] "Generic (PLEG): container finished" podID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerID="73c80c2c9d0d8cfaa3cca243004a745910379c837d85a4e6bce907ed5e68bf1c" exitCode=2 Oct 01 16:18:13 crc kubenswrapper[4726]: I1001 16:18:13.888456 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"547c9bca-03ba-4d79-8c13-c8506ce27252","Type":"ContainerDied","Data":"afd47f4eca931d1883629f439f47d59026063730435b0d205c5e545731a91049"} Oct 01 16:18:13 crc kubenswrapper[4726]: I1001 16:18:13.888726 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"547c9bca-03ba-4d79-8c13-c8506ce27252","Type":"ContainerDied","Data":"73c80c2c9d0d8cfaa3cca243004a745910379c837d85a4e6bce907ed5e68bf1c"} Oct 01 16:18:13 crc kubenswrapper[4726]: I1001 16:18:13.892399 4726 generic.go:334] "Generic (PLEG): container finished" podID="e86b00b2-1546-4cfc-a51a-23b92e950bc8" containerID="40826bad473aa6e993f5c08de182c0148854cccd39f29b519a8823126524c6ea" exitCode=0 Oct 01 16:18:13 crc kubenswrapper[4726]: I1001 16:18:13.892442 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e86b00b2-1546-4cfc-a51a-23b92e950bc8","Type":"ContainerDied","Data":"40826bad473aa6e993f5c08de182c0148854cccd39f29b519a8823126524c6ea"} Oct 01 16:18:14 crc kubenswrapper[4726]: I1001 16:18:14.912748 4726 generic.go:334] "Generic (PLEG): container finished" podID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerID="5c2a0c28fb1998f178aec0b143da14133c36382949f778b4035b4bb686b5d029" exitCode=0 Oct 01 16:18:14 crc kubenswrapper[4726]: I1001 16:18:14.912788 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"547c9bca-03ba-4d79-8c13-c8506ce27252","Type":"ContainerDied","Data":"5c2a0c28fb1998f178aec0b143da14133c36382949f778b4035b4bb686b5d029"} Oct 01 16:18:15 crc kubenswrapper[4726]: I1001 16:18:15.929080 4726 generic.go:334] "Generic (PLEG): container finished" podID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerID="4afa9e937c19dec8329428804da5e638a30d95fb07f42650d2de6a12ddec0eab" exitCode=0 Oct 01 16:18:15 crc kubenswrapper[4726]: I1001 16:18:15.929120 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"547c9bca-03ba-4d79-8c13-c8506ce27252","Type":"ContainerDied","Data":"4afa9e937c19dec8329428804da5e638a30d95fb07f42650d2de6a12ddec0eab"} Oct 01 16:18:16 crc kubenswrapper[4726]: E1001 16:18:16.233109 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod547c9bca_03ba_4d79_8c13_c8506ce27252.slice/crio-5c2a0c28fb1998f178aec0b143da14133c36382949f778b4035b4bb686b5d029.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod547c9bca_03ba_4d79_8c13_c8506ce27252.slice/crio-conmon-afd47f4eca931d1883629f439f47d59026063730435b0d205c5e545731a91049.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b0861ea_bb92_4592_9129_f9b3b4725a51.slice/crio-conmon-e870f4b47f8f55154e36b3b3cddea13fdb03c3983010f39eb6492d34869cb160.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34e6cb54_5cda_4ac4_bfe1_e59986daa349.slice/crio-bbb9783fb8558f7a4742d4e955d0800fcdeec69656503992055bbba90628b288\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod547c9bca_03ba_4d79_8c13_c8506ce27252.slice/crio-afd47f4eca931d1883629f439f47d59026063730435b0d205c5e545731a91049.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod547c9bca_03ba_4d79_8c13_c8506ce27252.slice/crio-73c80c2c9d0d8cfaa3cca243004a745910379c837d85a4e6bce907ed5e68bf1c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod547c9bca_03ba_4d79_8c13_c8506ce27252.slice/crio-conmon-5c2a0c28fb1998f178aec0b143da14133c36382949f778b4035b4bb686b5d029.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34e6cb54_5cda_4ac4_bfe1_e59986daa349.slice/crio-c82e46e2d4b5502528f154f2d29fabd0fe75b2be4ccdfcaee912aad6362e7885.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod547c9bca_03ba_4d79_8c13_c8506ce27252.slice/crio-conmon-73c80c2c9d0d8cfaa3cca243004a745910379c837d85a4e6bce907ed5e68bf1c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b0861ea_bb92_4592_9129_f9b3b4725a51.slice/crio-e870f4b47f8f55154e36b3b3cddea13fdb03c3983010f39eb6492d34869cb160.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode86b00b2_1546_4cfc_a51a_23b92e950bc8.slice/crio-conmon-40826bad473aa6e993f5c08de182c0148854cccd39f29b519a8823126524c6ea.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34e6cb54_5cda_4ac4_bfe1_e59986daa349.slice/crio-conmon-c82e46e2d4b5502528f154f2d29fabd0fe75b2be4ccdfcaee912aad6362e7885.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34e6cb54_5cda_4ac4_bfe1_e59986daa349.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode86b00b2_1546_4cfc_a51a_23b92e950bc8.slice/crio-conmon-857e05b3269cf36b182c4e6368942373520c3bc3f6759e90b5b9e98184750b1d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod547c9bca_03ba_4d79_8c13_c8506ce27252.slice/crio-4afa9e937c19dec8329428804da5e638a30d95fb07f42650d2de6a12ddec0eab.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode86b00b2_1546_4cfc_a51a_23b92e950bc8.slice/crio-40826bad473aa6e993f5c08de182c0148854cccd39f29b519a8823126524c6ea.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod547c9bca_03ba_4d79_8c13_c8506ce27252.slice/crio-conmon-4afa9e937c19dec8329428804da5e638a30d95fb07f42650d2de6a12ddec0eab.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode86b00b2_1546_4cfc_a51a_23b92e950bc8.slice/crio-857e05b3269cf36b182c4e6368942373520c3bc3f6759e90b5b9e98184750b1d.scope\": RecentStats: unable to find data in memory cache]" Oct 01 16:18:16 crc kubenswrapper[4726]: I1001 16:18:16.564779 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.154:3000/\": dial tcp 10.217.0.154:3000: connect: connection refused" Oct 01 16:18:16 crc kubenswrapper[4726]: I1001 16:18:16.925648 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 01 16:18:16 crc kubenswrapper[4726]: I1001 16:18:16.940389 4726 generic.go:334] "Generic (PLEG): container finished" podID="5b0861ea-bb92-4592-9129-f9b3b4725a51" containerID="e870f4b47f8f55154e36b3b3cddea13fdb03c3983010f39eb6492d34869cb160" exitCode=137 Oct 01 16:18:16 crc kubenswrapper[4726]: I1001 16:18:16.940442 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6ccb6664-v52zf" event={"ID":"5b0861ea-bb92-4592-9129-f9b3b4725a51","Type":"ContainerDied","Data":"e870f4b47f8f55154e36b3b3cddea13fdb03c3983010f39eb6492d34869cb160"} Oct 01 16:18:19 crc kubenswrapper[4726]: I1001 16:18:19.234825 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:19 crc kubenswrapper[4726]: I1001 16:18:19.735388 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5649c4d89d-cjbv9" Oct 01 16:18:19 crc kubenswrapper[4726]: I1001 16:18:19.835405 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7bc64455bb-r2zhn"] Oct 01 16:18:19 crc kubenswrapper[4726]: I1001 16:18:19.835926 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7bc64455bb-r2zhn" podUID="ebde1298-9550-4cac-a7a4-a4f31135b59f" containerName="barbican-api-log" containerID="cri-o://83fe40dbcf64c59c5ce1fbc0b8a6f2e82f3e7e48ddf8edd837c82f91d78058e1" gracePeriod=30 Oct 01 16:18:19 crc kubenswrapper[4726]: I1001 16:18:19.836111 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7bc64455bb-r2zhn" podUID="ebde1298-9550-4cac-a7a4-a4f31135b59f" containerName="barbican-api" containerID="cri-o://221cb6163476026452ea09ec2f2660001a03106ce1256f0cc5030dcb246a57dc" gracePeriod=30 Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.212212 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.227540 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.271615 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.297108 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-57c9dff847-vbzrr"] Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381115 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-scripts\") pod \"547c9bca-03ba-4d79-8c13-c8506ce27252\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381181 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b0861ea-bb92-4592-9129-f9b3b4725a51-logs\") pod \"5b0861ea-bb92-4592-9129-f9b3b4725a51\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381210 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b0861ea-bb92-4592-9129-f9b3b4725a51-scripts\") pod \"5b0861ea-bb92-4592-9129-f9b3b4725a51\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381248 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w88sr\" (UniqueName: \"kubernetes.io/projected/547c9bca-03ba-4d79-8c13-c8506ce27252-kube-api-access-w88sr\") pod \"547c9bca-03ba-4d79-8c13-c8506ce27252\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381290 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-combined-ca-bundle\") pod \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381327 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/547c9bca-03ba-4d79-8c13-c8506ce27252-run-httpd\") pod \"547c9bca-03ba-4d79-8c13-c8506ce27252\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381354 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-horizon-tls-certs\") pod \"5b0861ea-bb92-4592-9129-f9b3b4725a51\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381383 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx4z8\" (UniqueName: \"kubernetes.io/projected/e86b00b2-1546-4cfc-a51a-23b92e950bc8-kube-api-access-xx4z8\") pod \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381416 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-config-data\") pod \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381444 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e86b00b2-1546-4cfc-a51a-23b92e950bc8-etc-machine-id\") pod \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381526 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/547c9bca-03ba-4d79-8c13-c8506ce27252-log-httpd\") pod \"547c9bca-03ba-4d79-8c13-c8506ce27252\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381553 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-combined-ca-bundle\") pod \"5b0861ea-bb92-4592-9129-f9b3b4725a51\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381580 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-config-data-custom\") pod \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381624 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-config-data\") pod \"547c9bca-03ba-4d79-8c13-c8506ce27252\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381740 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-sg-core-conf-yaml\") pod \"547c9bca-03ba-4d79-8c13-c8506ce27252\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381773 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-combined-ca-bundle\") pod \"547c9bca-03ba-4d79-8c13-c8506ce27252\" (UID: \"547c9bca-03ba-4d79-8c13-c8506ce27252\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381801 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b0861ea-bb92-4592-9129-f9b3b4725a51-config-data\") pod \"5b0861ea-bb92-4592-9129-f9b3b4725a51\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381831 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztxl5\" (UniqueName: \"kubernetes.io/projected/5b0861ea-bb92-4592-9129-f9b3b4725a51-kube-api-access-ztxl5\") pod \"5b0861ea-bb92-4592-9129-f9b3b4725a51\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381854 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-scripts\") pod \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\" (UID: \"e86b00b2-1546-4cfc-a51a-23b92e950bc8\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.381879 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-horizon-secret-key\") pod \"5b0861ea-bb92-4592-9129-f9b3b4725a51\" (UID: \"5b0861ea-bb92-4592-9129-f9b3b4725a51\") " Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.382373 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e86b00b2-1546-4cfc-a51a-23b92e950bc8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e86b00b2-1546-4cfc-a51a-23b92e950bc8" (UID: "e86b00b2-1546-4cfc-a51a-23b92e950bc8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.385615 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/547c9bca-03ba-4d79-8c13-c8506ce27252-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "547c9bca-03ba-4d79-8c13-c8506ce27252" (UID: "547c9bca-03ba-4d79-8c13-c8506ce27252"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.395718 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/547c9bca-03ba-4d79-8c13-c8506ce27252-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "547c9bca-03ba-4d79-8c13-c8506ce27252" (UID: "547c9bca-03ba-4d79-8c13-c8506ce27252"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.400367 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b0861ea-bb92-4592-9129-f9b3b4725a51-logs" (OuterVolumeSpecName: "logs") pod "5b0861ea-bb92-4592-9129-f9b3b4725a51" (UID: "5b0861ea-bb92-4592-9129-f9b3b4725a51"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.413786 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/547c9bca-03ba-4d79-8c13-c8506ce27252-kube-api-access-w88sr" (OuterVolumeSpecName: "kube-api-access-w88sr") pod "547c9bca-03ba-4d79-8c13-c8506ce27252" (UID: "547c9bca-03ba-4d79-8c13-c8506ce27252"). InnerVolumeSpecName "kube-api-access-w88sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.421399 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e86b00b2-1546-4cfc-a51a-23b92e950bc8" (UID: "e86b00b2-1546-4cfc-a51a-23b92e950bc8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.427769 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-scripts" (OuterVolumeSpecName: "scripts") pod "547c9bca-03ba-4d79-8c13-c8506ce27252" (UID: "547c9bca-03ba-4d79-8c13-c8506ce27252"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.434969 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5b0861ea-bb92-4592-9129-f9b3b4725a51" (UID: "5b0861ea-bb92-4592-9129-f9b3b4725a51"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.435027 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b0861ea-bb92-4592-9129-f9b3b4725a51-kube-api-access-ztxl5" (OuterVolumeSpecName: "kube-api-access-ztxl5") pod "5b0861ea-bb92-4592-9129-f9b3b4725a51" (UID: "5b0861ea-bb92-4592-9129-f9b3b4725a51"). InnerVolumeSpecName "kube-api-access-ztxl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.456200 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-scripts" (OuterVolumeSpecName: "scripts") pod "e86b00b2-1546-4cfc-a51a-23b92e950bc8" (UID: "e86b00b2-1546-4cfc-a51a-23b92e950bc8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.476911 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e86b00b2-1546-4cfc-a51a-23b92e950bc8-kube-api-access-xx4z8" (OuterVolumeSpecName: "kube-api-access-xx4z8") pod "e86b00b2-1546-4cfc-a51a-23b92e950bc8" (UID: "e86b00b2-1546-4cfc-a51a-23b92e950bc8"). InnerVolumeSpecName "kube-api-access-xx4z8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.486582 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w88sr\" (UniqueName: \"kubernetes.io/projected/547c9bca-03ba-4d79-8c13-c8506ce27252-kube-api-access-w88sr\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.486618 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/547c9bca-03ba-4d79-8c13-c8506ce27252-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.486630 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx4z8\" (UniqueName: \"kubernetes.io/projected/e86b00b2-1546-4cfc-a51a-23b92e950bc8-kube-api-access-xx4z8\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.486642 4726 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e86b00b2-1546-4cfc-a51a-23b92e950bc8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.486655 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/547c9bca-03ba-4d79-8c13-c8506ce27252-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.486666 4726 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.486678 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztxl5\" (UniqueName: \"kubernetes.io/projected/5b0861ea-bb92-4592-9129-f9b3b4725a51-kube-api-access-ztxl5\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.486688 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.486700 4726 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.486711 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.486723 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b0861ea-bb92-4592-9129-f9b3b4725a51-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.491712 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b0861ea-bb92-4592-9129-f9b3b4725a51-scripts" (OuterVolumeSpecName: "scripts") pod "5b0861ea-bb92-4592-9129-f9b3b4725a51" (UID: "5b0861ea-bb92-4592-9129-f9b3b4725a51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.522143 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "547c9bca-03ba-4d79-8c13-c8506ce27252" (UID: "547c9bca-03ba-4d79-8c13-c8506ce27252"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.529150 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b0861ea-bb92-4592-9129-f9b3b4725a51-config-data" (OuterVolumeSpecName: "config-data") pod "5b0861ea-bb92-4592-9129-f9b3b4725a51" (UID: "5b0861ea-bb92-4592-9129-f9b3b4725a51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.536235 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b0861ea-bb92-4592-9129-f9b3b4725a51" (UID: "5b0861ea-bb92-4592-9129-f9b3b4725a51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.566370 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e86b00b2-1546-4cfc-a51a-23b92e950bc8" (UID: "e86b00b2-1546-4cfc-a51a-23b92e950bc8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.588280 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.588575 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.588654 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b0861ea-bb92-4592-9129-f9b3b4725a51-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.588725 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b0861ea-bb92-4592-9129-f9b3b4725a51-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.588791 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.628374 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "5b0861ea-bb92-4592-9129-f9b3b4725a51" (UID: "5b0861ea-bb92-4592-9129-f9b3b4725a51"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.632329 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "547c9bca-03ba-4d79-8c13-c8506ce27252" (UID: "547c9bca-03ba-4d79-8c13-c8506ce27252"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.651193 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-config-data" (OuterVolumeSpecName: "config-data") pod "547c9bca-03ba-4d79-8c13-c8506ce27252" (UID: "547c9bca-03ba-4d79-8c13-c8506ce27252"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.665311 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-config-data" (OuterVolumeSpecName: "config-data") pod "e86b00b2-1546-4cfc-a51a-23b92e950bc8" (UID: "e86b00b2-1546-4cfc-a51a-23b92e950bc8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.690507 4726 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b0861ea-bb92-4592-9129-f9b3b4725a51-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.690551 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86b00b2-1546-4cfc-a51a-23b92e950bc8-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.690564 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.690577 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547c9bca-03ba-4d79-8c13-c8506ce27252-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.981797 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"37c371f5-1756-49e1-82ec-40b7cb86f524","Type":"ContainerStarted","Data":"44cc20eb2ff6f7e5cf5bdd34505a742bd2095e71ef3974aca0856341ca300f5a"} Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.988924 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6ccb6664-v52zf" event={"ID":"5b0861ea-bb92-4592-9129-f9b3b4725a51","Type":"ContainerDied","Data":"3d525627067a1f7c05e703d40e50b8f2eb250c9ba0c9e7ae96f5bf8838a4bf31"} Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.988973 4726 scope.go:117] "RemoveContainer" containerID="fd9d715e9acaeb8f62267eb263a5863c5e412082eab1db4177fd27caa439f20a" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.989095 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c6ccb6664-v52zf" Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.995757 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-57c9dff847-vbzrr" event={"ID":"85ac310a-31b1-40c9-9724-2c488c0061e2","Type":"ContainerStarted","Data":"b3bc45836bcdf6253ab3f92c3251039cca3b4fad3c267135ca06726f80be89d6"} Oct 01 16:18:20 crc kubenswrapper[4726]: I1001 16:18:20.995809 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-57c9dff847-vbzrr" event={"ID":"85ac310a-31b1-40c9-9724-2c488c0061e2","Type":"ContainerStarted","Data":"ffbd1603432f38f05779f5a59fe85802ba9616846deabaaf576041f3b99c0e00"} Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.007194 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.007341 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e86b00b2-1546-4cfc-a51a-23b92e950bc8","Type":"ContainerDied","Data":"26e26efac83a73711399717f3eebc7c95f8b3f47742d885efa2b4b5d5d3eb62f"} Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.008496 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.465445561 podStartE2EDuration="18.008472204s" podCreationTimestamp="2025-10-01 16:18:03 +0000 UTC" firstStartedPulling="2025-10-01 16:18:04.710213847 +0000 UTC m=+1137.611766424" lastFinishedPulling="2025-10-01 16:18:20.25324049 +0000 UTC m=+1153.154793067" observedRunningTime="2025-10-01 16:18:21.000357122 +0000 UTC m=+1153.901909709" watchObservedRunningTime="2025-10-01 16:18:21.008472204 +0000 UTC m=+1153.910024791" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.022414 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.022410 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"547c9bca-03ba-4d79-8c13-c8506ce27252","Type":"ContainerDied","Data":"151710af01bedbf2ef9f8d74fe9aaa4e8df6eb5fb1f917c459a93cb1539f4978"} Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.027834 4726 generic.go:334] "Generic (PLEG): container finished" podID="ebde1298-9550-4cac-a7a4-a4f31135b59f" containerID="83fe40dbcf64c59c5ce1fbc0b8a6f2e82f3e7e48ddf8edd837c82f91d78058e1" exitCode=143 Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.027899 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc64455bb-r2zhn" event={"ID":"ebde1298-9550-4cac-a7a4-a4f31135b59f","Type":"ContainerDied","Data":"83fe40dbcf64c59c5ce1fbc0b8a6f2e82f3e7e48ddf8edd837c82f91d78058e1"} Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.110253 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c6ccb6664-v52zf"] Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.131264 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-c6ccb6664-v52zf"] Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.169937 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.178653 4726 scope.go:117] "RemoveContainer" containerID="e870f4b47f8f55154e36b3b3cddea13fdb03c3983010f39eb6492d34869cb160" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.187194 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.199289 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.204068 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.205412 4726 scope.go:117] "RemoveContainer" containerID="857e05b3269cf36b182c4e6368942373520c3bc3f6759e90b5b9e98184750b1d" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.210934 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:18:21 crc kubenswrapper[4726]: E1001 16:18:21.211359 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="ceilometer-central-agent" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211389 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="ceilometer-central-agent" Oct 01 16:18:21 crc kubenswrapper[4726]: E1001 16:18:21.211401 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="proxy-httpd" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211407 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="proxy-httpd" Oct 01 16:18:21 crc kubenswrapper[4726]: E1001 16:18:21.211419 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="ceilometer-notification-agent" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211425 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="ceilometer-notification-agent" Oct 01 16:18:21 crc kubenswrapper[4726]: E1001 16:18:21.211436 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0861ea-bb92-4592-9129-f9b3b4725a51" containerName="horizon" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211441 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0861ea-bb92-4592-9129-f9b3b4725a51" containerName="horizon" Oct 01 16:18:21 crc kubenswrapper[4726]: E1001 16:18:21.211461 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="sg-core" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211467 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="sg-core" Oct 01 16:18:21 crc kubenswrapper[4726]: E1001 16:18:21.211480 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e86b00b2-1546-4cfc-a51a-23b92e950bc8" containerName="cinder-scheduler" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211493 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e86b00b2-1546-4cfc-a51a-23b92e950bc8" containerName="cinder-scheduler" Oct 01 16:18:21 crc kubenswrapper[4726]: E1001 16:18:21.211506 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0861ea-bb92-4592-9129-f9b3b4725a51" containerName="horizon-log" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211512 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0861ea-bb92-4592-9129-f9b3b4725a51" containerName="horizon-log" Oct 01 16:18:21 crc kubenswrapper[4726]: E1001 16:18:21.211528 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e86b00b2-1546-4cfc-a51a-23b92e950bc8" containerName="probe" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211534 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e86b00b2-1546-4cfc-a51a-23b92e950bc8" containerName="probe" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211702 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0861ea-bb92-4592-9129-f9b3b4725a51" containerName="horizon-log" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211715 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="proxy-httpd" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211726 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="sg-core" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211740 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="ceilometer-notification-agent" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211750 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0861ea-bb92-4592-9129-f9b3b4725a51" containerName="horizon" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211757 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" containerName="ceilometer-central-agent" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211768 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e86b00b2-1546-4cfc-a51a-23b92e950bc8" containerName="probe" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.211777 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e86b00b2-1546-4cfc-a51a-23b92e950bc8" containerName="cinder-scheduler" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.212748 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.219754 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.219853 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.223756 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.228654 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.228830 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.232107 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.244186 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.244764 4726 scope.go:117] "RemoveContainer" containerID="40826bad473aa6e993f5c08de182c0148854cccd39f29b519a8823126524c6ea" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.271568 4726 scope.go:117] "RemoveContainer" containerID="afd47f4eca931d1883629f439f47d59026063730435b0d205c5e545731a91049" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.302438 4726 scope.go:117] "RemoveContainer" containerID="73c80c2c9d0d8cfaa3cca243004a745910379c837d85a4e6bce907ed5e68bf1c" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.314124 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae75086-c497-417c-82aa-9ed0e320b325-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.314359 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae75086-c497-417c-82aa-9ed0e320b325-scripts\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.314481 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aae75086-c497-417c-82aa-9ed0e320b325-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.315579 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrsw2\" (UniqueName: \"kubernetes.io/projected/aae75086-c497-417c-82aa-9ed0e320b325-kube-api-access-rrsw2\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.315718 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae75086-c497-417c-82aa-9ed0e320b325-config-data\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.315885 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aae75086-c497-417c-82aa-9ed0e320b325-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.319969 4726 scope.go:117] "RemoveContainer" containerID="4afa9e937c19dec8329428804da5e638a30d95fb07f42650d2de6a12ddec0eab" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.403038 4726 scope.go:117] "RemoveContainer" containerID="5c2a0c28fb1998f178aec0b143da14133c36382949f778b4035b4bb686b5d029" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.418087 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.418142 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvht6\" (UniqueName: \"kubernetes.io/projected/0e43b171-fe4c-4061-a92f-889493fb2e5d-kube-api-access-wvht6\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.418176 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aae75086-c497-417c-82aa-9ed0e320b325-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.418208 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-scripts\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.418246 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae75086-c497-417c-82aa-9ed0e320b325-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.418277 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae75086-c497-417c-82aa-9ed0e320b325-scripts\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.418310 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e43b171-fe4c-4061-a92f-889493fb2e5d-run-httpd\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.418344 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aae75086-c497-417c-82aa-9ed0e320b325-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.418383 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.418410 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrsw2\" (UniqueName: \"kubernetes.io/projected/aae75086-c497-417c-82aa-9ed0e320b325-kube-api-access-rrsw2\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.418450 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e43b171-fe4c-4061-a92f-889493fb2e5d-log-httpd\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.418502 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae75086-c497-417c-82aa-9ed0e320b325-config-data\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.418574 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-config-data\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.418709 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aae75086-c497-417c-82aa-9ed0e320b325-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.423793 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae75086-c497-417c-82aa-9ed0e320b325-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.427531 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae75086-c497-417c-82aa-9ed0e320b325-config-data\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.428037 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aae75086-c497-417c-82aa-9ed0e320b325-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.432447 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae75086-c497-417c-82aa-9ed0e320b325-scripts\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.447587 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrsw2\" (UniqueName: \"kubernetes.io/projected/aae75086-c497-417c-82aa-9ed0e320b325-kube-api-access-rrsw2\") pod \"cinder-scheduler-0\" (UID: \"aae75086-c497-417c-82aa-9ed0e320b325\") " pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.520202 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-config-data\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.520523 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.520545 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvht6\" (UniqueName: \"kubernetes.io/projected/0e43b171-fe4c-4061-a92f-889493fb2e5d-kube-api-access-wvht6\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.520569 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-scripts\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.520618 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e43b171-fe4c-4061-a92f-889493fb2e5d-run-httpd\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.520652 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.520686 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e43b171-fe4c-4061-a92f-889493fb2e5d-log-httpd\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.521117 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e43b171-fe4c-4061-a92f-889493fb2e5d-log-httpd\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.521309 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e43b171-fe4c-4061-a92f-889493fb2e5d-run-httpd\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.526602 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-config-data\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.527027 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.529515 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-scripts\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.530385 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.541395 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvht6\" (UniqueName: \"kubernetes.io/projected/0e43b171-fe4c-4061-a92f-889493fb2e5d-kube-api-access-wvht6\") pod \"ceilometer-0\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.543467 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.554658 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.821650 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="547c9bca-03ba-4d79-8c13-c8506ce27252" path="/var/lib/kubelet/pods/547c9bca-03ba-4d79-8c13-c8506ce27252/volumes" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.822403 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b0861ea-bb92-4592-9129-f9b3b4725a51" path="/var/lib/kubelet/pods/5b0861ea-bb92-4592-9129-f9b3b4725a51/volumes" Oct 01 16:18:21 crc kubenswrapper[4726]: I1001 16:18:21.823356 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e86b00b2-1546-4cfc-a51a-23b92e950bc8" path="/var/lib/kubelet/pods/e86b00b2-1546-4cfc-a51a-23b92e950bc8/volumes" Oct 01 16:18:22 crc kubenswrapper[4726]: I1001 16:18:22.053471 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-57c9dff847-vbzrr" event={"ID":"85ac310a-31b1-40c9-9724-2c488c0061e2","Type":"ContainerStarted","Data":"8a0cc0a66047d7d5303a520973e51effc62ea4a5569c80c15bbb310e7b0f26db"} Oct 01 16:18:22 crc kubenswrapper[4726]: I1001 16:18:22.053792 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:22 crc kubenswrapper[4726]: I1001 16:18:22.053806 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:22 crc kubenswrapper[4726]: W1001 16:18:22.081651 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e43b171_fe4c_4061_a92f_889493fb2e5d.slice/crio-203879a1d597add75f000a74b97e5a66eb3fed4299432d97b57f9a5ad75fdd6d WatchSource:0}: Error finding container 203879a1d597add75f000a74b97e5a66eb3fed4299432d97b57f9a5ad75fdd6d: Status 404 returned error can't find the container with id 203879a1d597add75f000a74b97e5a66eb3fed4299432d97b57f9a5ad75fdd6d Oct 01 16:18:22 crc kubenswrapper[4726]: I1001 16:18:22.081676 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:22 crc kubenswrapper[4726]: I1001 16:18:22.090276 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-57c9dff847-vbzrr" podStartSLOduration=10.090256226 podStartE2EDuration="10.090256226s" podCreationTimestamp="2025-10-01 16:18:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:18:22.087459006 +0000 UTC m=+1154.989011583" watchObservedRunningTime="2025-10-01 16:18:22.090256226 +0000 UTC m=+1154.991808803" Oct 01 16:18:22 crc kubenswrapper[4726]: W1001 16:18:22.149269 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaae75086_c497_417c_82aa_9ed0e320b325.slice/crio-8afc9caf7f6e42eb175a8c9c102af1b11137af0b65dc2e897af37f0324ca7c5c WatchSource:0}: Error finding container 8afc9caf7f6e42eb175a8c9c102af1b11137af0b65dc2e897af37f0324ca7c5c: Status 404 returned error can't find the container with id 8afc9caf7f6e42eb175a8c9c102af1b11137af0b65dc2e897af37f0324ca7c5c Oct 01 16:18:22 crc kubenswrapper[4726]: I1001 16:18:22.152407 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:18:22 crc kubenswrapper[4726]: I1001 16:18:22.888465 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:18:22 crc kubenswrapper[4726]: I1001 16:18:22.888890 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4d4b770c-28b0-47c1-ba00-584e3faa804c" containerName="glance-log" containerID="cri-o://bd33518a068228fb54a715e346710f1d194a7173acb87056b88a9e850b11315a" gracePeriod=30 Oct 01 16:18:22 crc kubenswrapper[4726]: I1001 16:18:22.888994 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4d4b770c-28b0-47c1-ba00-584e3faa804c" containerName="glance-httpd" containerID="cri-o://a26e79b359150c8cdd5d683f8211f3209fc5a21fbb8c7b2213808f40d87c20b3" gracePeriod=30 Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.011650 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7bc64455bb-r2zhn" podUID="ebde1298-9550-4cac-a7a4-a4f31135b59f" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:47962->10.217.0.158:9311: read: connection reset by peer" Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.011668 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7bc64455bb-r2zhn" podUID="ebde1298-9550-4cac-a7a4-a4f31135b59f" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:47950->10.217.0.158:9311: read: connection reset by peer" Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.067609 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"aae75086-c497-417c-82aa-9ed0e320b325","Type":"ContainerStarted","Data":"e57797cf2af5b5f548d15d935ca087fe8ddf2db68ea2d2962830ea46bdddda5c"} Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.067661 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"aae75086-c497-417c-82aa-9ed0e320b325","Type":"ContainerStarted","Data":"8afc9caf7f6e42eb175a8c9c102af1b11137af0b65dc2e897af37f0324ca7c5c"} Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.076972 4726 generic.go:334] "Generic (PLEG): container finished" podID="ebde1298-9550-4cac-a7a4-a4f31135b59f" containerID="221cb6163476026452ea09ec2f2660001a03106ce1256f0cc5030dcb246a57dc" exitCode=0 Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.077069 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc64455bb-r2zhn" event={"ID":"ebde1298-9550-4cac-a7a4-a4f31135b59f","Type":"ContainerDied","Data":"221cb6163476026452ea09ec2f2660001a03106ce1256f0cc5030dcb246a57dc"} Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.082027 4726 generic.go:334] "Generic (PLEG): container finished" podID="4d4b770c-28b0-47c1-ba00-584e3faa804c" containerID="bd33518a068228fb54a715e346710f1d194a7173acb87056b88a9e850b11315a" exitCode=143 Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.082122 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4d4b770c-28b0-47c1-ba00-584e3faa804c","Type":"ContainerDied","Data":"bd33518a068228fb54a715e346710f1d194a7173acb87056b88a9e850b11315a"} Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.090468 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e43b171-fe4c-4061-a92f-889493fb2e5d","Type":"ContainerStarted","Data":"958fad93c409c7a7e087f76753800626b023e2e9712e12043f75cc4c39241a17"} Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.090534 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e43b171-fe4c-4061-a92f-889493fb2e5d","Type":"ContainerStarted","Data":"203879a1d597add75f000a74b97e5a66eb3fed4299432d97b57f9a5ad75fdd6d"} Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.414663 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.414997 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.460442 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.569973 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebde1298-9550-4cac-a7a4-a4f31135b59f-logs\") pod \"ebde1298-9550-4cac-a7a4-a4f31135b59f\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.570071 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fh54\" (UniqueName: \"kubernetes.io/projected/ebde1298-9550-4cac-a7a4-a4f31135b59f-kube-api-access-6fh54\") pod \"ebde1298-9550-4cac-a7a4-a4f31135b59f\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.570257 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-config-data-custom\") pod \"ebde1298-9550-4cac-a7a4-a4f31135b59f\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.571083 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebde1298-9550-4cac-a7a4-a4f31135b59f-logs" (OuterVolumeSpecName: "logs") pod "ebde1298-9550-4cac-a7a4-a4f31135b59f" (UID: "ebde1298-9550-4cac-a7a4-a4f31135b59f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.571837 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-config-data\") pod \"ebde1298-9550-4cac-a7a4-a4f31135b59f\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.572212 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-combined-ca-bundle\") pod \"ebde1298-9550-4cac-a7a4-a4f31135b59f\" (UID: \"ebde1298-9550-4cac-a7a4-a4f31135b59f\") " Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.573032 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebde1298-9550-4cac-a7a4-a4f31135b59f-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.581197 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebde1298-9550-4cac-a7a4-a4f31135b59f-kube-api-access-6fh54" (OuterVolumeSpecName: "kube-api-access-6fh54") pod "ebde1298-9550-4cac-a7a4-a4f31135b59f" (UID: "ebde1298-9550-4cac-a7a4-a4f31135b59f"). InnerVolumeSpecName "kube-api-access-6fh54". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.583290 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ebde1298-9550-4cac-a7a4-a4f31135b59f" (UID: "ebde1298-9550-4cac-a7a4-a4f31135b59f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.615919 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ebde1298-9550-4cac-a7a4-a4f31135b59f" (UID: "ebde1298-9550-4cac-a7a4-a4f31135b59f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.665185 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-config-data" (OuterVolumeSpecName: "config-data") pod "ebde1298-9550-4cac-a7a4-a4f31135b59f" (UID: "ebde1298-9550-4cac-a7a4-a4f31135b59f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.674371 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fh54\" (UniqueName: \"kubernetes.io/projected/ebde1298-9550-4cac-a7a4-a4f31135b59f-kube-api-access-6fh54\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.674400 4726 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.674410 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:23 crc kubenswrapper[4726]: I1001 16:18:23.674418 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebde1298-9550-4cac-a7a4-a4f31135b59f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:24 crc kubenswrapper[4726]: I1001 16:18:24.099517 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc64455bb-r2zhn" event={"ID":"ebde1298-9550-4cac-a7a4-a4f31135b59f","Type":"ContainerDied","Data":"3fa7b320db45fb191c5c961a07f7e980b14bc8e1a564ce922d490fdea9e02b69"} Oct 01 16:18:24 crc kubenswrapper[4726]: I1001 16:18:24.099742 4726 scope.go:117] "RemoveContainer" containerID="221cb6163476026452ea09ec2f2660001a03106ce1256f0cc5030dcb246a57dc" Oct 01 16:18:24 crc kubenswrapper[4726]: I1001 16:18:24.099836 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bc64455bb-r2zhn" Oct 01 16:18:24 crc kubenswrapper[4726]: I1001 16:18:24.104589 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e43b171-fe4c-4061-a92f-889493fb2e5d","Type":"ContainerStarted","Data":"ee48c81fdbe1ca24c1e9d61501f44e6efbd5576a22c0c2c4bb1da39e58b89a34"} Oct 01 16:18:24 crc kubenswrapper[4726]: I1001 16:18:24.107357 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"aae75086-c497-417c-82aa-9ed0e320b325","Type":"ContainerStarted","Data":"245482205fd45a60b12e54a1ea2c12d8847f26244ef952584b771365f0736614"} Oct 01 16:18:24 crc kubenswrapper[4726]: I1001 16:18:24.123385 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7bc64455bb-r2zhn"] Oct 01 16:18:24 crc kubenswrapper[4726]: I1001 16:18:24.128861 4726 scope.go:117] "RemoveContainer" containerID="83fe40dbcf64c59c5ce1fbc0b8a6f2e82f3e7e48ddf8edd837c82f91d78058e1" Oct 01 16:18:24 crc kubenswrapper[4726]: I1001 16:18:24.132367 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7bc64455bb-r2zhn"] Oct 01 16:18:24 crc kubenswrapper[4726]: I1001 16:18:24.135136 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.13511536 podStartE2EDuration="3.13511536s" podCreationTimestamp="2025-10-01 16:18:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:18:24.132802484 +0000 UTC m=+1157.034355061" watchObservedRunningTime="2025-10-01 16:18:24.13511536 +0000 UTC m=+1157.036667937" Oct 01 16:18:25 crc kubenswrapper[4726]: I1001 16:18:25.119095 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e43b171-fe4c-4061-a92f-889493fb2e5d","Type":"ContainerStarted","Data":"9cfc4504a0948f0156844abdb0681fc4051c68e8f4b2e9dfea22b94c27d6ef41"} Oct 01 16:18:25 crc kubenswrapper[4726]: I1001 16:18:25.765356 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:18:25 crc kubenswrapper[4726]: I1001 16:18:25.765584 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c4283abe-f876-41c5-aa74-348a7396091f" containerName="glance-log" containerID="cri-o://c2aa66ea86e3e5a10e1d812a1c8294b844da3d725de401714373e9be0e2200a1" gracePeriod=30 Oct 01 16:18:25 crc kubenswrapper[4726]: I1001 16:18:25.765715 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c4283abe-f876-41c5-aa74-348a7396091f" containerName="glance-httpd" containerID="cri-o://cfbcc613ec702e703d65c0eb38d0837c1cede2e51f043c403845485988b616ca" gracePeriod=30 Oct 01 16:18:25 crc kubenswrapper[4726]: I1001 16:18:25.826094 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebde1298-9550-4cac-a7a4-a4f31135b59f" path="/var/lib/kubelet/pods/ebde1298-9550-4cac-a7a4-a4f31135b59f/volumes" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.130093 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e43b171-fe4c-4061-a92f-889493fb2e5d","Type":"ContainerStarted","Data":"f05ffca2c0e4fe6194c1aeebfebfd4c60c6e275b7a0a092f2aeaa63a598abdc4"} Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.131368 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.133968 4726 generic.go:334] "Generic (PLEG): container finished" podID="c4283abe-f876-41c5-aa74-348a7396091f" containerID="c2aa66ea86e3e5a10e1d812a1c8294b844da3d725de401714373e9be0e2200a1" exitCode=143 Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.134021 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c4283abe-f876-41c5-aa74-348a7396091f","Type":"ContainerDied","Data":"c2aa66ea86e3e5a10e1d812a1c8294b844da3d725de401714373e9be0e2200a1"} Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.139560 4726 generic.go:334] "Generic (PLEG): container finished" podID="4d4b770c-28b0-47c1-ba00-584e3faa804c" containerID="a26e79b359150c8cdd5d683f8211f3209fc5a21fbb8c7b2213808f40d87c20b3" exitCode=0 Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.139599 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4d4b770c-28b0-47c1-ba00-584e3faa804c","Type":"ContainerDied","Data":"a26e79b359150c8cdd5d683f8211f3209fc5a21fbb8c7b2213808f40d87c20b3"} Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.160684 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.558870072 podStartE2EDuration="5.160662021s" podCreationTimestamp="2025-10-01 16:18:21 +0000 UTC" firstStartedPulling="2025-10-01 16:18:22.089094503 +0000 UTC m=+1154.990647090" lastFinishedPulling="2025-10-01 16:18:25.690886442 +0000 UTC m=+1158.592439039" observedRunningTime="2025-10-01 16:18:26.150017607 +0000 UTC m=+1159.051570204" watchObservedRunningTime="2025-10-01 16:18:26.160662021 +0000 UTC m=+1159.062214598" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.341951 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-8v9wq"] Oct 01 16:18:26 crc kubenswrapper[4726]: E1001 16:18:26.342562 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebde1298-9550-4cac-a7a4-a4f31135b59f" containerName="barbican-api" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.342578 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebde1298-9550-4cac-a7a4-a4f31135b59f" containerName="barbican-api" Oct 01 16:18:26 crc kubenswrapper[4726]: E1001 16:18:26.342603 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebde1298-9550-4cac-a7a4-a4f31135b59f" containerName="barbican-api-log" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.342609 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebde1298-9550-4cac-a7a4-a4f31135b59f" containerName="barbican-api-log" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.342784 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebde1298-9550-4cac-a7a4-a4f31135b59f" containerName="barbican-api-log" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.342804 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebde1298-9550-4cac-a7a4-a4f31135b59f" containerName="barbican-api" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.346551 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8v9wq" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.365764 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8v9wq"] Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.445858 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-hvk5c"] Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.446923 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hvk5c" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.468095 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-hvk5c"] Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.529082 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6mg7\" (UniqueName: \"kubernetes.io/projected/1ff89aa8-5b64-461c-9cad-7fe1e0375c77-kube-api-access-k6mg7\") pod \"nova-api-db-create-8v9wq\" (UID: \"1ff89aa8-5b64-461c-9cad-7fe1e0375c77\") " pod="openstack/nova-api-db-create-8v9wq" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.546411 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.575085 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-kcw4f"] Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.577119 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kcw4f" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.590887 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kcw4f"] Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.638989 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6mg7\" (UniqueName: \"kubernetes.io/projected/1ff89aa8-5b64-461c-9cad-7fe1e0375c77-kube-api-access-k6mg7\") pod \"nova-api-db-create-8v9wq\" (UID: \"1ff89aa8-5b64-461c-9cad-7fe1e0375c77\") " pod="openstack/nova-api-db-create-8v9wq" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.639099 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz987\" (UniqueName: \"kubernetes.io/projected/9d0b9d80-050e-47bf-a71d-1ca4b10ac868-kube-api-access-nz987\") pod \"nova-cell0-db-create-hvk5c\" (UID: \"9d0b9d80-050e-47bf-a71d-1ca4b10ac868\") " pod="openstack/nova-cell0-db-create-hvk5c" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.650268 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.656630 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6mg7\" (UniqueName: \"kubernetes.io/projected/1ff89aa8-5b64-461c-9cad-7fe1e0375c77-kube-api-access-k6mg7\") pod \"nova-api-db-create-8v9wq\" (UID: \"1ff89aa8-5b64-461c-9cad-7fe1e0375c77\") " pod="openstack/nova-api-db-create-8v9wq" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.711255 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8v9wq" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.719098 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.740414 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz987\" (UniqueName: \"kubernetes.io/projected/9d0b9d80-050e-47bf-a71d-1ca4b10ac868-kube-api-access-nz987\") pod \"nova-cell0-db-create-hvk5c\" (UID: \"9d0b9d80-050e-47bf-a71d-1ca4b10ac868\") " pod="openstack/nova-cell0-db-create-hvk5c" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.740578 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mmdz\" (UniqueName: \"kubernetes.io/projected/c5eadc81-68d4-44be-acd8-6127518dcbdf-kube-api-access-8mmdz\") pod \"nova-cell1-db-create-kcw4f\" (UID: \"c5eadc81-68d4-44be-acd8-6127518dcbdf\") " pod="openstack/nova-cell1-db-create-kcw4f" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.761775 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz987\" (UniqueName: \"kubernetes.io/projected/9d0b9d80-050e-47bf-a71d-1ca4b10ac868-kube-api-access-nz987\") pod \"nova-cell0-db-create-hvk5c\" (UID: \"9d0b9d80-050e-47bf-a71d-1ca4b10ac868\") " pod="openstack/nova-cell0-db-create-hvk5c" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.773480 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hvk5c" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.845542 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d4b770c-28b0-47c1-ba00-584e3faa804c-logs\") pod \"4d4b770c-28b0-47c1-ba00-584e3faa804c\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.845644 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-scripts\") pod \"4d4b770c-28b0-47c1-ba00-584e3faa804c\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.845691 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-combined-ca-bundle\") pod \"4d4b770c-28b0-47c1-ba00-584e3faa804c\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.845755 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-public-tls-certs\") pod \"4d4b770c-28b0-47c1-ba00-584e3faa804c\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.845817 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4d4b770c-28b0-47c1-ba00-584e3faa804c-httpd-run\") pod \"4d4b770c-28b0-47c1-ba00-584e3faa804c\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.845864 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-config-data\") pod \"4d4b770c-28b0-47c1-ba00-584e3faa804c\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.845892 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"4d4b770c-28b0-47c1-ba00-584e3faa804c\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.845912 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2sm7\" (UniqueName: \"kubernetes.io/projected/4d4b770c-28b0-47c1-ba00-584e3faa804c-kube-api-access-s2sm7\") pod \"4d4b770c-28b0-47c1-ba00-584e3faa804c\" (UID: \"4d4b770c-28b0-47c1-ba00-584e3faa804c\") " Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.846640 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mmdz\" (UniqueName: \"kubernetes.io/projected/c5eadc81-68d4-44be-acd8-6127518dcbdf-kube-api-access-8mmdz\") pod \"nova-cell1-db-create-kcw4f\" (UID: \"c5eadc81-68d4-44be-acd8-6127518dcbdf\") " pod="openstack/nova-cell1-db-create-kcw4f" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.848415 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d4b770c-28b0-47c1-ba00-584e3faa804c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4d4b770c-28b0-47c1-ba00-584e3faa804c" (UID: "4d4b770c-28b0-47c1-ba00-584e3faa804c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.848639 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d4b770c-28b0-47c1-ba00-584e3faa804c-logs" (OuterVolumeSpecName: "logs") pod "4d4b770c-28b0-47c1-ba00-584e3faa804c" (UID: "4d4b770c-28b0-47c1-ba00-584e3faa804c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.854247 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d4b770c-28b0-47c1-ba00-584e3faa804c-kube-api-access-s2sm7" (OuterVolumeSpecName: "kube-api-access-s2sm7") pod "4d4b770c-28b0-47c1-ba00-584e3faa804c" (UID: "4d4b770c-28b0-47c1-ba00-584e3faa804c"). InnerVolumeSpecName "kube-api-access-s2sm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.858189 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "4d4b770c-28b0-47c1-ba00-584e3faa804c" (UID: "4d4b770c-28b0-47c1-ba00-584e3faa804c"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.869237 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-scripts" (OuterVolumeSpecName: "scripts") pod "4d4b770c-28b0-47c1-ba00-584e3faa804c" (UID: "4d4b770c-28b0-47c1-ba00-584e3faa804c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.873666 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mmdz\" (UniqueName: \"kubernetes.io/projected/c5eadc81-68d4-44be-acd8-6127518dcbdf-kube-api-access-8mmdz\") pod \"nova-cell1-db-create-kcw4f\" (UID: \"c5eadc81-68d4-44be-acd8-6127518dcbdf\") " pod="openstack/nova-cell1-db-create-kcw4f" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.897790 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kcw4f" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.913445 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d4b770c-28b0-47c1-ba00-584e3faa804c" (UID: "4d4b770c-28b0-47c1-ba00-584e3faa804c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.932335 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4d4b770c-28b0-47c1-ba00-584e3faa804c" (UID: "4d4b770c-28b0-47c1-ba00-584e3faa804c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.948340 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.948368 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.948378 4726 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.948386 4726 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4d4b770c-28b0-47c1-ba00-584e3faa804c-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.948407 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.948418 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2sm7\" (UniqueName: \"kubernetes.io/projected/4d4b770c-28b0-47c1-ba00-584e3faa804c-kube-api-access-s2sm7\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.948427 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d4b770c-28b0-47c1-ba00-584e3faa804c-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.964220 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-config-data" (OuterVolumeSpecName: "config-data") pod "4d4b770c-28b0-47c1-ba00-584e3faa804c" (UID: "4d4b770c-28b0-47c1-ba00-584e3faa804c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:26 crc kubenswrapper[4726]: I1001 16:18:26.995282 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.050192 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4b770c-28b0-47c1-ba00-584e3faa804c-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.050217 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.156041 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4d4b770c-28b0-47c1-ba00-584e3faa804c","Type":"ContainerDied","Data":"5f2976fabdad8e6ede7b523c8a431f4ea34a3f85cf0e39a7e293a13c40bcb790"} Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.156131 4726 scope.go:117] "RemoveContainer" containerID="a26e79b359150c8cdd5d683f8211f3209fc5a21fbb8c7b2213808f40d87c20b3" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.156276 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.196411 4726 scope.go:117] "RemoveContainer" containerID="bd33518a068228fb54a715e346710f1d194a7173acb87056b88a9e850b11315a" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.201994 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.217440 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.239131 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:18:27 crc kubenswrapper[4726]: E1001 16:18:27.239909 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d4b770c-28b0-47c1-ba00-584e3faa804c" containerName="glance-log" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.240009 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d4b770c-28b0-47c1-ba00-584e3faa804c" containerName="glance-log" Oct 01 16:18:27 crc kubenswrapper[4726]: E1001 16:18:27.240178 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d4b770c-28b0-47c1-ba00-584e3faa804c" containerName="glance-httpd" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.240263 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d4b770c-28b0-47c1-ba00-584e3faa804c" containerName="glance-httpd" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.240579 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d4b770c-28b0-47c1-ba00-584e3faa804c" containerName="glance-log" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.240680 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d4b770c-28b0-47c1-ba00-584e3faa804c" containerName="glance-httpd" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.241825 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.250907 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.251841 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.259495 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.286500 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8v9wq"] Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.351892 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-hvk5c"] Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.354995 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.355115 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7aa87400-a85d-4d53-8b65-915138d87c01-scripts\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.355206 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7aa87400-a85d-4d53-8b65-915138d87c01-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.355270 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7aa87400-a85d-4d53-8b65-915138d87c01-logs\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.355457 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7aa87400-a85d-4d53-8b65-915138d87c01-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.355603 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aa87400-a85d-4d53-8b65-915138d87c01-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.355652 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4crrj\" (UniqueName: \"kubernetes.io/projected/7aa87400-a85d-4d53-8b65-915138d87c01-kube-api-access-4crrj\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.355823 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa87400-a85d-4d53-8b65-915138d87c01-config-data\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.425324 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kcw4f"] Oct 01 16:18:27 crc kubenswrapper[4726]: W1001 16:18:27.432186 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5eadc81_68d4_44be_acd8_6127518dcbdf.slice/crio-15baa406eae6c6025bfe02b919dca83ce66f2a2dce9b8d182f6365d65483b6a7 WatchSource:0}: Error finding container 15baa406eae6c6025bfe02b919dca83ce66f2a2dce9b8d182f6365d65483b6a7: Status 404 returned error can't find the container with id 15baa406eae6c6025bfe02b919dca83ce66f2a2dce9b8d182f6365d65483b6a7 Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.457234 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aa87400-a85d-4d53-8b65-915138d87c01-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.457294 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4crrj\" (UniqueName: \"kubernetes.io/projected/7aa87400-a85d-4d53-8b65-915138d87c01-kube-api-access-4crrj\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.457379 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa87400-a85d-4d53-8b65-915138d87c01-config-data\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.457438 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.457461 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7aa87400-a85d-4d53-8b65-915138d87c01-scripts\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.457482 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7aa87400-a85d-4d53-8b65-915138d87c01-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.457502 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7aa87400-a85d-4d53-8b65-915138d87c01-logs\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.457565 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7aa87400-a85d-4d53-8b65-915138d87c01-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.458016 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.458140 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7aa87400-a85d-4d53-8b65-915138d87c01-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.458263 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7aa87400-a85d-4d53-8b65-915138d87c01-logs\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.463783 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7aa87400-a85d-4d53-8b65-915138d87c01-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.466277 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa87400-a85d-4d53-8b65-915138d87c01-config-data\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.467010 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aa87400-a85d-4d53-8b65-915138d87c01-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.467574 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7aa87400-a85d-4d53-8b65-915138d87c01-scripts\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.477965 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4crrj\" (UniqueName: \"kubernetes.io/projected/7aa87400-a85d-4d53-8b65-915138d87c01-kube-api-access-4crrj\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.496824 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"7aa87400-a85d-4d53-8b65-915138d87c01\") " pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.583569 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.600362 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.607842 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-57c9dff847-vbzrr" Oct 01 16:18:27 crc kubenswrapper[4726]: I1001 16:18:27.828387 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d4b770c-28b0-47c1-ba00-584e3faa804c" path="/var/lib/kubelet/pods/4d4b770c-28b0-47c1-ba00-584e3faa804c/volumes" Oct 01 16:18:28 crc kubenswrapper[4726]: I1001 16:18:28.175788 4726 generic.go:334] "Generic (PLEG): container finished" podID="c5eadc81-68d4-44be-acd8-6127518dcbdf" containerID="de70527985aba717e6a1d83443850b0ccb7a8b9147bbc29c1458868ccebc1135" exitCode=0 Oct 01 16:18:28 crc kubenswrapper[4726]: I1001 16:18:28.175885 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kcw4f" event={"ID":"c5eadc81-68d4-44be-acd8-6127518dcbdf","Type":"ContainerDied","Data":"de70527985aba717e6a1d83443850b0ccb7a8b9147bbc29c1458868ccebc1135"} Oct 01 16:18:28 crc kubenswrapper[4726]: I1001 16:18:28.175917 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kcw4f" event={"ID":"c5eadc81-68d4-44be-acd8-6127518dcbdf","Type":"ContainerStarted","Data":"15baa406eae6c6025bfe02b919dca83ce66f2a2dce9b8d182f6365d65483b6a7"} Oct 01 16:18:28 crc kubenswrapper[4726]: I1001 16:18:28.190846 4726 generic.go:334] "Generic (PLEG): container finished" podID="1ff89aa8-5b64-461c-9cad-7fe1e0375c77" containerID="5487ced454167d5e1713318c6c6a40500aa65b70bce9aade694640a0a5973e99" exitCode=0 Oct 01 16:18:28 crc kubenswrapper[4726]: I1001 16:18:28.191352 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8v9wq" event={"ID":"1ff89aa8-5b64-461c-9cad-7fe1e0375c77","Type":"ContainerDied","Data":"5487ced454167d5e1713318c6c6a40500aa65b70bce9aade694640a0a5973e99"} Oct 01 16:18:28 crc kubenswrapper[4726]: I1001 16:18:28.191412 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8v9wq" event={"ID":"1ff89aa8-5b64-461c-9cad-7fe1e0375c77","Type":"ContainerStarted","Data":"fe3840080402588b9c627c7f12289c86bce86e40281daed39ce7787c67cb9e61"} Oct 01 16:18:28 crc kubenswrapper[4726]: I1001 16:18:28.208803 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:18:28 crc kubenswrapper[4726]: I1001 16:18:28.211962 4726 generic.go:334] "Generic (PLEG): container finished" podID="9d0b9d80-050e-47bf-a71d-1ca4b10ac868" containerID="b329989b95b7dfdbec6f920c028e2c191fcaedf3425895056eaf276a1e2f3c2a" exitCode=0 Oct 01 16:18:28 crc kubenswrapper[4726]: I1001 16:18:28.212529 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hvk5c" event={"ID":"9d0b9d80-050e-47bf-a71d-1ca4b10ac868","Type":"ContainerDied","Data":"b329989b95b7dfdbec6f920c028e2c191fcaedf3425895056eaf276a1e2f3c2a"} Oct 01 16:18:28 crc kubenswrapper[4726]: I1001 16:18:28.212568 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="proxy-httpd" containerID="cri-o://f05ffca2c0e4fe6194c1aeebfebfd4c60c6e275b7a0a092f2aeaa63a598abdc4" gracePeriod=30 Oct 01 16:18:28 crc kubenswrapper[4726]: I1001 16:18:28.212585 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hvk5c" event={"ID":"9d0b9d80-050e-47bf-a71d-1ca4b10ac868","Type":"ContainerStarted","Data":"c3a90b58f1abd209197c908e20f1a91e5f71c1e00e65293bfd03d397fd722e22"} Oct 01 16:18:28 crc kubenswrapper[4726]: I1001 16:18:28.212707 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="sg-core" containerID="cri-o://9cfc4504a0948f0156844abdb0681fc4051c68e8f4b2e9dfea22b94c27d6ef41" gracePeriod=30 Oct 01 16:18:28 crc kubenswrapper[4726]: I1001 16:18:28.212773 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="ceilometer-notification-agent" containerID="cri-o://ee48c81fdbe1ca24c1e9d61501f44e6efbd5576a22c0c2c4bb1da39e58b89a34" gracePeriod=30 Oct 01 16:18:28 crc kubenswrapper[4726]: I1001 16:18:28.212824 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="ceilometer-central-agent" containerID="cri-o://958fad93c409c7a7e087f76753800626b023e2e9712e12043f75cc4c39241a17" gracePeriod=30 Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.031654 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.233447 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7aa87400-a85d-4d53-8b65-915138d87c01","Type":"ContainerStarted","Data":"84c13259e22274b59a315a3aac066ccc7b001b8c66fffd3fdcff25a400d88cd9"} Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.233733 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7aa87400-a85d-4d53-8b65-915138d87c01","Type":"ContainerStarted","Data":"6c01218e9a7dd4f93a7c1e86f50ce63ea2f7c97c08c6f508f33b5a94a6b5134b"} Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.243066 4726 generic.go:334] "Generic (PLEG): container finished" podID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerID="f05ffca2c0e4fe6194c1aeebfebfd4c60c6e275b7a0a092f2aeaa63a598abdc4" exitCode=0 Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.243092 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e43b171-fe4c-4061-a92f-889493fb2e5d","Type":"ContainerDied","Data":"f05ffca2c0e4fe6194c1aeebfebfd4c60c6e275b7a0a092f2aeaa63a598abdc4"} Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.243157 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e43b171-fe4c-4061-a92f-889493fb2e5d","Type":"ContainerDied","Data":"9cfc4504a0948f0156844abdb0681fc4051c68e8f4b2e9dfea22b94c27d6ef41"} Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.243122 4726 generic.go:334] "Generic (PLEG): container finished" podID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerID="9cfc4504a0948f0156844abdb0681fc4051c68e8f4b2e9dfea22b94c27d6ef41" exitCode=2 Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.243182 4726 generic.go:334] "Generic (PLEG): container finished" podID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerID="ee48c81fdbe1ca24c1e9d61501f44e6efbd5576a22c0c2c4bb1da39e58b89a34" exitCode=0 Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.243283 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e43b171-fe4c-4061-a92f-889493fb2e5d","Type":"ContainerDied","Data":"ee48c81fdbe1ca24c1e9d61501f44e6efbd5576a22c0c2c4bb1da39e58b89a34"} Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.248137 4726 generic.go:334] "Generic (PLEG): container finished" podID="c4283abe-f876-41c5-aa74-348a7396091f" containerID="cfbcc613ec702e703d65c0eb38d0837c1cede2e51f043c403845485988b616ca" exitCode=0 Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.248174 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c4283abe-f876-41c5-aa74-348a7396091f","Type":"ContainerDied","Data":"cfbcc613ec702e703d65c0eb38d0837c1cede2e51f043c403845485988b616ca"} Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.433264 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.600283 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-config-data\") pod \"c4283abe-f876-41c5-aa74-348a7396091f\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.600344 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-internal-tls-certs\") pod \"c4283abe-f876-41c5-aa74-348a7396091f\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.600414 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c4283abe-f876-41c5-aa74-348a7396091f-httpd-run\") pod \"c4283abe-f876-41c5-aa74-348a7396091f\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.600444 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-scripts\") pod \"c4283abe-f876-41c5-aa74-348a7396091f\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.600482 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltk56\" (UniqueName: \"kubernetes.io/projected/c4283abe-f876-41c5-aa74-348a7396091f-kube-api-access-ltk56\") pod \"c4283abe-f876-41c5-aa74-348a7396091f\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.600506 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"c4283abe-f876-41c5-aa74-348a7396091f\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.600588 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4283abe-f876-41c5-aa74-348a7396091f-logs\") pod \"c4283abe-f876-41c5-aa74-348a7396091f\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.600619 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-combined-ca-bundle\") pod \"c4283abe-f876-41c5-aa74-348a7396091f\" (UID: \"c4283abe-f876-41c5-aa74-348a7396091f\") " Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.608914 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4283abe-f876-41c5-aa74-348a7396091f-logs" (OuterVolumeSpecName: "logs") pod "c4283abe-f876-41c5-aa74-348a7396091f" (UID: "c4283abe-f876-41c5-aa74-348a7396091f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.609048 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4283abe-f876-41c5-aa74-348a7396091f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c4283abe-f876-41c5-aa74-348a7396091f" (UID: "c4283abe-f876-41c5-aa74-348a7396091f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.610159 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-scripts" (OuterVolumeSpecName: "scripts") pod "c4283abe-f876-41c5-aa74-348a7396091f" (UID: "c4283abe-f876-41c5-aa74-348a7396091f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.612223 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4283abe-f876-41c5-aa74-348a7396091f-kube-api-access-ltk56" (OuterVolumeSpecName: "kube-api-access-ltk56") pod "c4283abe-f876-41c5-aa74-348a7396091f" (UID: "c4283abe-f876-41c5-aa74-348a7396091f"). InnerVolumeSpecName "kube-api-access-ltk56". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.616213 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "c4283abe-f876-41c5-aa74-348a7396091f" (UID: "c4283abe-f876-41c5-aa74-348a7396091f"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.632732 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4283abe-f876-41c5-aa74-348a7396091f" (UID: "c4283abe-f876-41c5-aa74-348a7396091f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.677404 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c4283abe-f876-41c5-aa74-348a7396091f" (UID: "c4283abe-f876-41c5-aa74-348a7396091f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.682459 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kcw4f" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.702666 4726 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.702695 4726 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c4283abe-f876-41c5-aa74-348a7396091f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.702704 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.702715 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltk56\" (UniqueName: \"kubernetes.io/projected/c4283abe-f876-41c5-aa74-348a7396091f-kube-api-access-ltk56\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.702737 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.702746 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4283abe-f876-41c5-aa74-348a7396091f-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.702753 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.716307 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-config-data" (OuterVolumeSpecName: "config-data") pod "c4283abe-f876-41c5-aa74-348a7396091f" (UID: "c4283abe-f876-41c5-aa74-348a7396091f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.723907 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.803677 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mmdz\" (UniqueName: \"kubernetes.io/projected/c5eadc81-68d4-44be-acd8-6127518dcbdf-kube-api-access-8mmdz\") pod \"c5eadc81-68d4-44be-acd8-6127518dcbdf\" (UID: \"c5eadc81-68d4-44be-acd8-6127518dcbdf\") " Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.804214 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4283abe-f876-41c5-aa74-348a7396091f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.804249 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.809223 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5eadc81-68d4-44be-acd8-6127518dcbdf-kube-api-access-8mmdz" (OuterVolumeSpecName: "kube-api-access-8mmdz") pod "c5eadc81-68d4-44be-acd8-6127518dcbdf" (UID: "c5eadc81-68d4-44be-acd8-6127518dcbdf"). InnerVolumeSpecName "kube-api-access-8mmdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.884124 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hvk5c" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.891421 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8v9wq" Oct 01 16:18:29 crc kubenswrapper[4726]: I1001 16:18:29.908201 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mmdz\" (UniqueName: \"kubernetes.io/projected/c5eadc81-68d4-44be-acd8-6127518dcbdf-kube-api-access-8mmdz\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.009160 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz987\" (UniqueName: \"kubernetes.io/projected/9d0b9d80-050e-47bf-a71d-1ca4b10ac868-kube-api-access-nz987\") pod \"9d0b9d80-050e-47bf-a71d-1ca4b10ac868\" (UID: \"9d0b9d80-050e-47bf-a71d-1ca4b10ac868\") " Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.009332 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6mg7\" (UniqueName: \"kubernetes.io/projected/1ff89aa8-5b64-461c-9cad-7fe1e0375c77-kube-api-access-k6mg7\") pod \"1ff89aa8-5b64-461c-9cad-7fe1e0375c77\" (UID: \"1ff89aa8-5b64-461c-9cad-7fe1e0375c77\") " Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.015991 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d0b9d80-050e-47bf-a71d-1ca4b10ac868-kube-api-access-nz987" (OuterVolumeSpecName: "kube-api-access-nz987") pod "9d0b9d80-050e-47bf-a71d-1ca4b10ac868" (UID: "9d0b9d80-050e-47bf-a71d-1ca4b10ac868"). InnerVolumeSpecName "kube-api-access-nz987". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.016234 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ff89aa8-5b64-461c-9cad-7fe1e0375c77-kube-api-access-k6mg7" (OuterVolumeSpecName: "kube-api-access-k6mg7") pod "1ff89aa8-5b64-461c-9cad-7fe1e0375c77" (UID: "1ff89aa8-5b64-461c-9cad-7fe1e0375c77"). InnerVolumeSpecName "kube-api-access-k6mg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.110972 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6mg7\" (UniqueName: \"kubernetes.io/projected/1ff89aa8-5b64-461c-9cad-7fe1e0375c77-kube-api-access-k6mg7\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.111002 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz987\" (UniqueName: \"kubernetes.io/projected/9d0b9d80-050e-47bf-a71d-1ca4b10ac868-kube-api-access-nz987\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.256936 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hvk5c" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.260182 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hvk5c" event={"ID":"9d0b9d80-050e-47bf-a71d-1ca4b10ac868","Type":"ContainerDied","Data":"c3a90b58f1abd209197c908e20f1a91e5f71c1e00e65293bfd03d397fd722e22"} Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.260235 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3a90b58f1abd209197c908e20f1a91e5f71c1e00e65293bfd03d397fd722e22" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.262320 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7aa87400-a85d-4d53-8b65-915138d87c01","Type":"ContainerStarted","Data":"7134b0e56b97312c7fe304a1d69b3e37e9a2df76fdda5861ebcc0f852205ec83"} Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.264775 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.264770 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c4283abe-f876-41c5-aa74-348a7396091f","Type":"ContainerDied","Data":"b446ee5c74757b35755688e1eee3b7122b75ca3a1cc4c1acfff8ecb0ac8d22c2"} Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.264901 4726 scope.go:117] "RemoveContainer" containerID="cfbcc613ec702e703d65c0eb38d0837c1cede2e51f043c403845485988b616ca" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.266370 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kcw4f" event={"ID":"c5eadc81-68d4-44be-acd8-6127518dcbdf","Type":"ContainerDied","Data":"15baa406eae6c6025bfe02b919dca83ce66f2a2dce9b8d182f6365d65483b6a7"} Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.266398 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15baa406eae6c6025bfe02b919dca83ce66f2a2dce9b8d182f6365d65483b6a7" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.266444 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kcw4f" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.267635 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8v9wq" event={"ID":"1ff89aa8-5b64-461c-9cad-7fe1e0375c77","Type":"ContainerDied","Data":"fe3840080402588b9c627c7f12289c86bce86e40281daed39ce7787c67cb9e61"} Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.267899 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8v9wq" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.267917 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe3840080402588b9c627c7f12289c86bce86e40281daed39ce7787c67cb9e61" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.296414 4726 scope.go:117] "RemoveContainer" containerID="c2aa66ea86e3e5a10e1d812a1c8294b844da3d725de401714373e9be0e2200a1" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.297656 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.297643929 podStartE2EDuration="3.297643929s" podCreationTimestamp="2025-10-01 16:18:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:18:30.294772157 +0000 UTC m=+1163.196324764" watchObservedRunningTime="2025-10-01 16:18:30.297643929 +0000 UTC m=+1163.199196506" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.324606 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.340953 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.356343 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:18:30 crc kubenswrapper[4726]: E1001 16:18:30.356735 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4283abe-f876-41c5-aa74-348a7396091f" containerName="glance-log" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.356751 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4283abe-f876-41c5-aa74-348a7396091f" containerName="glance-log" Oct 01 16:18:30 crc kubenswrapper[4726]: E1001 16:18:30.356765 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d0b9d80-050e-47bf-a71d-1ca4b10ac868" containerName="mariadb-database-create" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.356772 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d0b9d80-050e-47bf-a71d-1ca4b10ac868" containerName="mariadb-database-create" Oct 01 16:18:30 crc kubenswrapper[4726]: E1001 16:18:30.356795 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ff89aa8-5b64-461c-9cad-7fe1e0375c77" containerName="mariadb-database-create" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.356802 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ff89aa8-5b64-461c-9cad-7fe1e0375c77" containerName="mariadb-database-create" Oct 01 16:18:30 crc kubenswrapper[4726]: E1001 16:18:30.356817 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5eadc81-68d4-44be-acd8-6127518dcbdf" containerName="mariadb-database-create" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.356825 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5eadc81-68d4-44be-acd8-6127518dcbdf" containerName="mariadb-database-create" Oct 01 16:18:30 crc kubenswrapper[4726]: E1001 16:18:30.356842 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4283abe-f876-41c5-aa74-348a7396091f" containerName="glance-httpd" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.356850 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4283abe-f876-41c5-aa74-348a7396091f" containerName="glance-httpd" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.357043 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4283abe-f876-41c5-aa74-348a7396091f" containerName="glance-log" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.357092 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5eadc81-68d4-44be-acd8-6127518dcbdf" containerName="mariadb-database-create" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.357113 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d0b9d80-050e-47bf-a71d-1ca4b10ac868" containerName="mariadb-database-create" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.357129 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4283abe-f876-41c5-aa74-348a7396091f" containerName="glance-httpd" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.357139 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ff89aa8-5b64-461c-9cad-7fe1e0375c77" containerName="mariadb-database-create" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.358034 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.360219 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.375460 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.397608 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.517642 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4a301ba-9e78-4262-9efa-3d4d91a5957a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.517711 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4a301ba-9e78-4262-9efa-3d4d91a5957a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.517744 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cvhm\" (UniqueName: \"kubernetes.io/projected/b4a301ba-9e78-4262-9efa-3d4d91a5957a-kube-api-access-2cvhm\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.517791 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4a301ba-9e78-4262-9efa-3d4d91a5957a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.517819 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4a301ba-9e78-4262-9efa-3d4d91a5957a-logs\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.517865 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.517900 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4a301ba-9e78-4262-9efa-3d4d91a5957a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.517928 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4a301ba-9e78-4262-9efa-3d4d91a5957a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.619119 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4a301ba-9e78-4262-9efa-3d4d91a5957a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.619186 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4a301ba-9e78-4262-9efa-3d4d91a5957a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.619222 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cvhm\" (UniqueName: \"kubernetes.io/projected/b4a301ba-9e78-4262-9efa-3d4d91a5957a-kube-api-access-2cvhm\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.619257 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4a301ba-9e78-4262-9efa-3d4d91a5957a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.619290 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4a301ba-9e78-4262-9efa-3d4d91a5957a-logs\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.619348 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.619384 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4a301ba-9e78-4262-9efa-3d4d91a5957a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.619411 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4a301ba-9e78-4262-9efa-3d4d91a5957a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.619871 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.619959 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4a301ba-9e78-4262-9efa-3d4d91a5957a-logs\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.620274 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4a301ba-9e78-4262-9efa-3d4d91a5957a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.624999 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4a301ba-9e78-4262-9efa-3d4d91a5957a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.625134 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4a301ba-9e78-4262-9efa-3d4d91a5957a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.625919 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4a301ba-9e78-4262-9efa-3d4d91a5957a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.626216 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4a301ba-9e78-4262-9efa-3d4d91a5957a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.640287 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cvhm\" (UniqueName: \"kubernetes.io/projected/b4a301ba-9e78-4262-9efa-3d4d91a5957a-kube-api-access-2cvhm\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.651231 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"b4a301ba-9e78-4262-9efa-3d4d91a5957a\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:18:30 crc kubenswrapper[4726]: I1001 16:18:30.689179 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:18:31 crc kubenswrapper[4726]: I1001 16:18:31.249209 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:18:31 crc kubenswrapper[4726]: I1001 16:18:31.278041 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b4a301ba-9e78-4262-9efa-3d4d91a5957a","Type":"ContainerStarted","Data":"ba96af07171637c34ac6f90146d35bcfc62256d3863d0c08a9b9fb9da7892e25"} Oct 01 16:18:31 crc kubenswrapper[4726]: I1001 16:18:31.801643 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 01 16:18:31 crc kubenswrapper[4726]: I1001 16:18:31.819784 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4283abe-f876-41c5-aa74-348a7396091f" path="/var/lib/kubelet/pods/c4283abe-f876-41c5-aa74-348a7396091f/volumes" Oct 01 16:18:32 crc kubenswrapper[4726]: I1001 16:18:32.293443 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b4a301ba-9e78-4262-9efa-3d4d91a5957a","Type":"ContainerStarted","Data":"41c2a2c92ba90e3954d21c917a966c4f28165080b2e36c9ec40ad3fbe2b3f488"} Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.266565 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.303773 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b4a301ba-9e78-4262-9efa-3d4d91a5957a","Type":"ContainerStarted","Data":"c89262ec2f8286459262ede5ffb0b6e86acabd8bbd33bc72d2ef26252b8644e8"} Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.307473 4726 generic.go:334] "Generic (PLEG): container finished" podID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerID="958fad93c409c7a7e087f76753800626b023e2e9712e12043f75cc4c39241a17" exitCode=0 Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.307514 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e43b171-fe4c-4061-a92f-889493fb2e5d","Type":"ContainerDied","Data":"958fad93c409c7a7e087f76753800626b023e2e9712e12043f75cc4c39241a17"} Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.307539 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e43b171-fe4c-4061-a92f-889493fb2e5d","Type":"ContainerDied","Data":"203879a1d597add75f000a74b97e5a66eb3fed4299432d97b57f9a5ad75fdd6d"} Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.307542 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.307557 4726 scope.go:117] "RemoveContainer" containerID="f05ffca2c0e4fe6194c1aeebfebfd4c60c6e275b7a0a092f2aeaa63a598abdc4" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.338907 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.338884525 podStartE2EDuration="3.338884525s" podCreationTimestamp="2025-10-01 16:18:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:18:33.327603813 +0000 UTC m=+1166.229156430" watchObservedRunningTime="2025-10-01 16:18:33.338884525 +0000 UTC m=+1166.240437152" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.340084 4726 scope.go:117] "RemoveContainer" containerID="9cfc4504a0948f0156844abdb0681fc4051c68e8f4b2e9dfea22b94c27d6ef41" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.359068 4726 scope.go:117] "RemoveContainer" containerID="ee48c81fdbe1ca24c1e9d61501f44e6efbd5576a22c0c2c4bb1da39e58b89a34" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.365059 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-config-data\") pod \"0e43b171-fe4c-4061-a92f-889493fb2e5d\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.365088 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-scripts\") pod \"0e43b171-fe4c-4061-a92f-889493fb2e5d\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.365192 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e43b171-fe4c-4061-a92f-889493fb2e5d-log-httpd\") pod \"0e43b171-fe4c-4061-a92f-889493fb2e5d\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.365250 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e43b171-fe4c-4061-a92f-889493fb2e5d-run-httpd\") pod \"0e43b171-fe4c-4061-a92f-889493fb2e5d\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.365282 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvht6\" (UniqueName: \"kubernetes.io/projected/0e43b171-fe4c-4061-a92f-889493fb2e5d-kube-api-access-wvht6\") pod \"0e43b171-fe4c-4061-a92f-889493fb2e5d\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.365390 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-combined-ca-bundle\") pod \"0e43b171-fe4c-4061-a92f-889493fb2e5d\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.365453 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-sg-core-conf-yaml\") pod \"0e43b171-fe4c-4061-a92f-889493fb2e5d\" (UID: \"0e43b171-fe4c-4061-a92f-889493fb2e5d\") " Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.366364 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e43b171-fe4c-4061-a92f-889493fb2e5d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0e43b171-fe4c-4061-a92f-889493fb2e5d" (UID: "0e43b171-fe4c-4061-a92f-889493fb2e5d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.366374 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e43b171-fe4c-4061-a92f-889493fb2e5d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0e43b171-fe4c-4061-a92f-889493fb2e5d" (UID: "0e43b171-fe4c-4061-a92f-889493fb2e5d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.385926 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-scripts" (OuterVolumeSpecName: "scripts") pod "0e43b171-fe4c-4061-a92f-889493fb2e5d" (UID: "0e43b171-fe4c-4061-a92f-889493fb2e5d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.386094 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e43b171-fe4c-4061-a92f-889493fb2e5d-kube-api-access-wvht6" (OuterVolumeSpecName: "kube-api-access-wvht6") pod "0e43b171-fe4c-4061-a92f-889493fb2e5d" (UID: "0e43b171-fe4c-4061-a92f-889493fb2e5d"). InnerVolumeSpecName "kube-api-access-wvht6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.386296 4726 scope.go:117] "RemoveContainer" containerID="958fad93c409c7a7e087f76753800626b023e2e9712e12043f75cc4c39241a17" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.392937 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0e43b171-fe4c-4061-a92f-889493fb2e5d" (UID: "0e43b171-fe4c-4061-a92f-889493fb2e5d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.460744 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e43b171-fe4c-4061-a92f-889493fb2e5d" (UID: "0e43b171-fe4c-4061-a92f-889493fb2e5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.467124 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-config-data" (OuterVolumeSpecName: "config-data") pod "0e43b171-fe4c-4061-a92f-889493fb2e5d" (UID: "0e43b171-fe4c-4061-a92f-889493fb2e5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.467876 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e43b171-fe4c-4061-a92f-889493fb2e5d-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.467899 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e43b171-fe4c-4061-a92f-889493fb2e5d-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.467910 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvht6\" (UniqueName: \"kubernetes.io/projected/0e43b171-fe4c-4061-a92f-889493fb2e5d-kube-api-access-wvht6\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.467920 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.467929 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.467936 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.467943 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e43b171-fe4c-4061-a92f-889493fb2e5d-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.497452 4726 scope.go:117] "RemoveContainer" containerID="f05ffca2c0e4fe6194c1aeebfebfd4c60c6e275b7a0a092f2aeaa63a598abdc4" Oct 01 16:18:33 crc kubenswrapper[4726]: E1001 16:18:33.499484 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f05ffca2c0e4fe6194c1aeebfebfd4c60c6e275b7a0a092f2aeaa63a598abdc4\": container with ID starting with f05ffca2c0e4fe6194c1aeebfebfd4c60c6e275b7a0a092f2aeaa63a598abdc4 not found: ID does not exist" containerID="f05ffca2c0e4fe6194c1aeebfebfd4c60c6e275b7a0a092f2aeaa63a598abdc4" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.499524 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f05ffca2c0e4fe6194c1aeebfebfd4c60c6e275b7a0a092f2aeaa63a598abdc4"} err="failed to get container status \"f05ffca2c0e4fe6194c1aeebfebfd4c60c6e275b7a0a092f2aeaa63a598abdc4\": rpc error: code = NotFound desc = could not find container \"f05ffca2c0e4fe6194c1aeebfebfd4c60c6e275b7a0a092f2aeaa63a598abdc4\": container with ID starting with f05ffca2c0e4fe6194c1aeebfebfd4c60c6e275b7a0a092f2aeaa63a598abdc4 not found: ID does not exist" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.499555 4726 scope.go:117] "RemoveContainer" containerID="9cfc4504a0948f0156844abdb0681fc4051c68e8f4b2e9dfea22b94c27d6ef41" Oct 01 16:18:33 crc kubenswrapper[4726]: E1001 16:18:33.499891 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cfc4504a0948f0156844abdb0681fc4051c68e8f4b2e9dfea22b94c27d6ef41\": container with ID starting with 9cfc4504a0948f0156844abdb0681fc4051c68e8f4b2e9dfea22b94c27d6ef41 not found: ID does not exist" containerID="9cfc4504a0948f0156844abdb0681fc4051c68e8f4b2e9dfea22b94c27d6ef41" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.499929 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cfc4504a0948f0156844abdb0681fc4051c68e8f4b2e9dfea22b94c27d6ef41"} err="failed to get container status \"9cfc4504a0948f0156844abdb0681fc4051c68e8f4b2e9dfea22b94c27d6ef41\": rpc error: code = NotFound desc = could not find container \"9cfc4504a0948f0156844abdb0681fc4051c68e8f4b2e9dfea22b94c27d6ef41\": container with ID starting with 9cfc4504a0948f0156844abdb0681fc4051c68e8f4b2e9dfea22b94c27d6ef41 not found: ID does not exist" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.499955 4726 scope.go:117] "RemoveContainer" containerID="ee48c81fdbe1ca24c1e9d61501f44e6efbd5576a22c0c2c4bb1da39e58b89a34" Oct 01 16:18:33 crc kubenswrapper[4726]: E1001 16:18:33.500365 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee48c81fdbe1ca24c1e9d61501f44e6efbd5576a22c0c2c4bb1da39e58b89a34\": container with ID starting with ee48c81fdbe1ca24c1e9d61501f44e6efbd5576a22c0c2c4bb1da39e58b89a34 not found: ID does not exist" containerID="ee48c81fdbe1ca24c1e9d61501f44e6efbd5576a22c0c2c4bb1da39e58b89a34" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.500390 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee48c81fdbe1ca24c1e9d61501f44e6efbd5576a22c0c2c4bb1da39e58b89a34"} err="failed to get container status \"ee48c81fdbe1ca24c1e9d61501f44e6efbd5576a22c0c2c4bb1da39e58b89a34\": rpc error: code = NotFound desc = could not find container \"ee48c81fdbe1ca24c1e9d61501f44e6efbd5576a22c0c2c4bb1da39e58b89a34\": container with ID starting with ee48c81fdbe1ca24c1e9d61501f44e6efbd5576a22c0c2c4bb1da39e58b89a34 not found: ID does not exist" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.500404 4726 scope.go:117] "RemoveContainer" containerID="958fad93c409c7a7e087f76753800626b023e2e9712e12043f75cc4c39241a17" Oct 01 16:18:33 crc kubenswrapper[4726]: E1001 16:18:33.500691 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"958fad93c409c7a7e087f76753800626b023e2e9712e12043f75cc4c39241a17\": container with ID starting with 958fad93c409c7a7e087f76753800626b023e2e9712e12043f75cc4c39241a17 not found: ID does not exist" containerID="958fad93c409c7a7e087f76753800626b023e2e9712e12043f75cc4c39241a17" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.500718 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"958fad93c409c7a7e087f76753800626b023e2e9712e12043f75cc4c39241a17"} err="failed to get container status \"958fad93c409c7a7e087f76753800626b023e2e9712e12043f75cc4c39241a17\": rpc error: code = NotFound desc = could not find container \"958fad93c409c7a7e087f76753800626b023e2e9712e12043f75cc4c39241a17\": container with ID starting with 958fad93c409c7a7e087f76753800626b023e2e9712e12043f75cc4c39241a17 not found: ID does not exist" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.668806 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.681642 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.692191 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:33 crc kubenswrapper[4726]: E1001 16:18:33.692679 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="ceilometer-notification-agent" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.692705 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="ceilometer-notification-agent" Oct 01 16:18:33 crc kubenswrapper[4726]: E1001 16:18:33.692733 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="ceilometer-central-agent" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.692742 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="ceilometer-central-agent" Oct 01 16:18:33 crc kubenswrapper[4726]: E1001 16:18:33.692767 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="sg-core" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.692776 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="sg-core" Oct 01 16:18:33 crc kubenswrapper[4726]: E1001 16:18:33.692786 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="proxy-httpd" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.692794 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="proxy-httpd" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.693031 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="proxy-httpd" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.693071 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="ceilometer-central-agent" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.693095 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="sg-core" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.693106 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" containerName="ceilometer-notification-agent" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.695107 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.702228 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.704607 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.713577 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.773362 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-scripts\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.773417 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-config-data\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.773472 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e65728b8-4153-40f8-ad98-c3b6abacf176-log-httpd\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.773499 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qjds\" (UniqueName: \"kubernetes.io/projected/e65728b8-4153-40f8-ad98-c3b6abacf176-kube-api-access-7qjds\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.773535 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.773596 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.773655 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e65728b8-4153-40f8-ad98-c3b6abacf176-run-httpd\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.821528 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e43b171-fe4c-4061-a92f-889493fb2e5d" path="/var/lib/kubelet/pods/0e43b171-fe4c-4061-a92f-889493fb2e5d/volumes" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.875976 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-scripts\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.876032 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-config-data\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.876889 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e65728b8-4153-40f8-ad98-c3b6abacf176-log-httpd\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.877206 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qjds\" (UniqueName: \"kubernetes.io/projected/e65728b8-4153-40f8-ad98-c3b6abacf176-kube-api-access-7qjds\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.877266 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.877342 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.877461 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e65728b8-4153-40f8-ad98-c3b6abacf176-run-httpd\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.878232 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e65728b8-4153-40f8-ad98-c3b6abacf176-log-httpd\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.879832 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-config-data\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.879891 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e65728b8-4153-40f8-ad98-c3b6abacf176-run-httpd\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.881684 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-scripts\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.882413 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.892648 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:33 crc kubenswrapper[4726]: I1001 16:18:33.896089 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qjds\" (UniqueName: \"kubernetes.io/projected/e65728b8-4153-40f8-ad98-c3b6abacf176-kube-api-access-7qjds\") pod \"ceilometer-0\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " pod="openstack/ceilometer-0" Oct 01 16:18:34 crc kubenswrapper[4726]: I1001 16:18:34.027014 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:34 crc kubenswrapper[4726]: I1001 16:18:34.485738 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:34 crc kubenswrapper[4726]: W1001 16:18:34.492149 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode65728b8_4153_40f8_ad98_c3b6abacf176.slice/crio-65b82d1fee7e578b3617d9ecf87e555563acb46c55ca394c0cd5d0f657ba299c WatchSource:0}: Error finding container 65b82d1fee7e578b3617d9ecf87e555563acb46c55ca394c0cd5d0f657ba299c: Status 404 returned error can't find the container with id 65b82d1fee7e578b3617d9ecf87e555563acb46c55ca394c0cd5d0f657ba299c Oct 01 16:18:34 crc kubenswrapper[4726]: I1001 16:18:34.635728 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-598fc759b9-28cln" Oct 01 16:18:34 crc kubenswrapper[4726]: I1001 16:18:34.646913 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:34 crc kubenswrapper[4726]: I1001 16:18:34.692963 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-566c76dfb4-x49km"] Oct 01 16:18:34 crc kubenswrapper[4726]: I1001 16:18:34.693263 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-566c76dfb4-x49km" podUID="e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" containerName="neutron-api" containerID="cri-o://bd05040824b5af9e2be2be07f72261af0136cfbbfca823ddfc3960a35db693bf" gracePeriod=30 Oct 01 16:18:34 crc kubenswrapper[4726]: I1001 16:18:34.693296 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-566c76dfb4-x49km" podUID="e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" containerName="neutron-httpd" containerID="cri-o://eae9e5f7a8a4413ed91ee774a7f3ec41fe23fab3f71b4aca60e097647f504fb5" gracePeriod=30 Oct 01 16:18:35 crc kubenswrapper[4726]: I1001 16:18:35.325636 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e65728b8-4153-40f8-ad98-c3b6abacf176","Type":"ContainerStarted","Data":"eb4a6ee6fafd7da92c626c4937c8b6ee39ffa9671dad6a6aee4baf659ad73063"} Oct 01 16:18:35 crc kubenswrapper[4726]: I1001 16:18:35.325956 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e65728b8-4153-40f8-ad98-c3b6abacf176","Type":"ContainerStarted","Data":"65b82d1fee7e578b3617d9ecf87e555563acb46c55ca394c0cd5d0f657ba299c"} Oct 01 16:18:35 crc kubenswrapper[4726]: I1001 16:18:35.327528 4726 generic.go:334] "Generic (PLEG): container finished" podID="e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" containerID="eae9e5f7a8a4413ed91ee774a7f3ec41fe23fab3f71b4aca60e097647f504fb5" exitCode=0 Oct 01 16:18:35 crc kubenswrapper[4726]: I1001 16:18:35.327556 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-566c76dfb4-x49km" event={"ID":"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984","Type":"ContainerDied","Data":"eae9e5f7a8a4413ed91ee774a7f3ec41fe23fab3f71b4aca60e097647f504fb5"} Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.483318 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-f3eb-account-create-6g7cm"] Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.484960 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f3eb-account-create-6g7cm" Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.487375 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.497891 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f3eb-account-create-6g7cm"] Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.529134 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdj6s\" (UniqueName: \"kubernetes.io/projected/a7a05e67-296b-4ed9-b63c-30c6edc152db-kube-api-access-kdj6s\") pod \"nova-api-f3eb-account-create-6g7cm\" (UID: \"a7a05e67-296b-4ed9-b63c-30c6edc152db\") " pod="openstack/nova-api-f3eb-account-create-6g7cm" Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.630760 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdj6s\" (UniqueName: \"kubernetes.io/projected/a7a05e67-296b-4ed9-b63c-30c6edc152db-kube-api-access-kdj6s\") pod \"nova-api-f3eb-account-create-6g7cm\" (UID: \"a7a05e67-296b-4ed9-b63c-30c6edc152db\") " pod="openstack/nova-api-f3eb-account-create-6g7cm" Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.658295 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdj6s\" (UniqueName: \"kubernetes.io/projected/a7a05e67-296b-4ed9-b63c-30c6edc152db-kube-api-access-kdj6s\") pod \"nova-api-f3eb-account-create-6g7cm\" (UID: \"a7a05e67-296b-4ed9-b63c-30c6edc152db\") " pod="openstack/nova-api-f3eb-account-create-6g7cm" Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.695252 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-560d-account-create-fztkr"] Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.696722 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-560d-account-create-fztkr" Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.698543 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.706216 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-560d-account-create-fztkr"] Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.732030 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhhpn\" (UniqueName: \"kubernetes.io/projected/1c7c4328-6015-4e9a-9cd8-67c1d275928e-kube-api-access-bhhpn\") pod \"nova-cell0-560d-account-create-fztkr\" (UID: \"1c7c4328-6015-4e9a-9cd8-67c1d275928e\") " pod="openstack/nova-cell0-560d-account-create-fztkr" Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.803591 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f3eb-account-create-6g7cm" Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.834708 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhhpn\" (UniqueName: \"kubernetes.io/projected/1c7c4328-6015-4e9a-9cd8-67c1d275928e-kube-api-access-bhhpn\") pod \"nova-cell0-560d-account-create-fztkr\" (UID: \"1c7c4328-6015-4e9a-9cd8-67c1d275928e\") " pod="openstack/nova-cell0-560d-account-create-fztkr" Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.853080 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhhpn\" (UniqueName: \"kubernetes.io/projected/1c7c4328-6015-4e9a-9cd8-67c1d275928e-kube-api-access-bhhpn\") pod \"nova-cell0-560d-account-create-fztkr\" (UID: \"1c7c4328-6015-4e9a-9cd8-67c1d275928e\") " pod="openstack/nova-cell0-560d-account-create-fztkr" Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.884342 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-b942-account-create-7x8nm"] Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.885866 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b942-account-create-7x8nm" Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.888005 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.891551 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b942-account-create-7x8nm"] Oct 01 16:18:36 crc kubenswrapper[4726]: I1001 16:18:36.940715 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nxpm\" (UniqueName: \"kubernetes.io/projected/82d29008-5715-4a39-b0ea-252402970cdc-kube-api-access-2nxpm\") pod \"nova-cell1-b942-account-create-7x8nm\" (UID: \"82d29008-5715-4a39-b0ea-252402970cdc\") " pod="openstack/nova-cell1-b942-account-create-7x8nm" Oct 01 16:18:37 crc kubenswrapper[4726]: I1001 16:18:37.018673 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-560d-account-create-fztkr" Oct 01 16:18:37 crc kubenswrapper[4726]: I1001 16:18:37.042684 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nxpm\" (UniqueName: \"kubernetes.io/projected/82d29008-5715-4a39-b0ea-252402970cdc-kube-api-access-2nxpm\") pod \"nova-cell1-b942-account-create-7x8nm\" (UID: \"82d29008-5715-4a39-b0ea-252402970cdc\") " pod="openstack/nova-cell1-b942-account-create-7x8nm" Oct 01 16:18:37 crc kubenswrapper[4726]: I1001 16:18:37.064137 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nxpm\" (UniqueName: \"kubernetes.io/projected/82d29008-5715-4a39-b0ea-252402970cdc-kube-api-access-2nxpm\") pod \"nova-cell1-b942-account-create-7x8nm\" (UID: \"82d29008-5715-4a39-b0ea-252402970cdc\") " pod="openstack/nova-cell1-b942-account-create-7x8nm" Oct 01 16:18:37 crc kubenswrapper[4726]: I1001 16:18:37.218482 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b942-account-create-7x8nm" Oct 01 16:18:37 crc kubenswrapper[4726]: I1001 16:18:37.285435 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f3eb-account-create-6g7cm"] Oct 01 16:18:37 crc kubenswrapper[4726]: I1001 16:18:37.352500 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f3eb-account-create-6g7cm" event={"ID":"a7a05e67-296b-4ed9-b63c-30c6edc152db","Type":"ContainerStarted","Data":"7007d0b56c16e5086399bc0089f2a1d31a296d20cecc12708e5bb33d6821729b"} Oct 01 16:18:37 crc kubenswrapper[4726]: I1001 16:18:37.354940 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e65728b8-4153-40f8-ad98-c3b6abacf176","Type":"ContainerStarted","Data":"472776a7552756900eed556bff8d94abd8bcc8c1c2f9ee3b7e2fd75376435097"} Oct 01 16:18:37 crc kubenswrapper[4726]: I1001 16:18:37.354970 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e65728b8-4153-40f8-ad98-c3b6abacf176","Type":"ContainerStarted","Data":"09d1c5c0629432eb0d489832b2dc004f400b8a2207d2409e1e2766c0795e3f23"} Oct 01 16:18:37 crc kubenswrapper[4726]: I1001 16:18:37.487698 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-560d-account-create-fztkr"] Oct 01 16:18:37 crc kubenswrapper[4726]: I1001 16:18:37.584695 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 16:18:37 crc kubenswrapper[4726]: I1001 16:18:37.584740 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 16:18:37 crc kubenswrapper[4726]: I1001 16:18:37.627016 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 16:18:37 crc kubenswrapper[4726]: I1001 16:18:37.661097 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 16:18:37 crc kubenswrapper[4726]: W1001 16:18:37.711834 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82d29008_5715_4a39_b0ea_252402970cdc.slice/crio-5dd9d30832bd1af9b339db3db0cfb39acaec0e1201154dc0b83beb39a3ee0212 WatchSource:0}: Error finding container 5dd9d30832bd1af9b339db3db0cfb39acaec0e1201154dc0b83beb39a3ee0212: Status 404 returned error can't find the container with id 5dd9d30832bd1af9b339db3db0cfb39acaec0e1201154dc0b83beb39a3ee0212 Oct 01 16:18:37 crc kubenswrapper[4726]: I1001 16:18:37.720788 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b942-account-create-7x8nm"] Oct 01 16:18:38 crc kubenswrapper[4726]: I1001 16:18:38.366320 4726 generic.go:334] "Generic (PLEG): container finished" podID="1c7c4328-6015-4e9a-9cd8-67c1d275928e" containerID="c501b4ff046e6ca2c04e937e150b7e1489a3ddce4f4c61c941e381e3f4f52b6a" exitCode=0 Oct 01 16:18:38 crc kubenswrapper[4726]: I1001 16:18:38.366404 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-560d-account-create-fztkr" event={"ID":"1c7c4328-6015-4e9a-9cd8-67c1d275928e","Type":"ContainerDied","Data":"c501b4ff046e6ca2c04e937e150b7e1489a3ddce4f4c61c941e381e3f4f52b6a"} Oct 01 16:18:38 crc kubenswrapper[4726]: I1001 16:18:38.368278 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-560d-account-create-fztkr" event={"ID":"1c7c4328-6015-4e9a-9cd8-67c1d275928e","Type":"ContainerStarted","Data":"504f67c9d9acf76055509e758242536957f01df42dc4627949c8b3a28394d992"} Oct 01 16:18:38 crc kubenswrapper[4726]: I1001 16:18:38.379718 4726 generic.go:334] "Generic (PLEG): container finished" podID="a7a05e67-296b-4ed9-b63c-30c6edc152db" containerID="2a92e8f1fe7f8f0e9de69da93cf026e59dcdbe7e99806622f897846e76c43954" exitCode=0 Oct 01 16:18:38 crc kubenswrapper[4726]: I1001 16:18:38.379812 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f3eb-account-create-6g7cm" event={"ID":"a7a05e67-296b-4ed9-b63c-30c6edc152db","Type":"ContainerDied","Data":"2a92e8f1fe7f8f0e9de69da93cf026e59dcdbe7e99806622f897846e76c43954"} Oct 01 16:18:38 crc kubenswrapper[4726]: I1001 16:18:38.382236 4726 generic.go:334] "Generic (PLEG): container finished" podID="82d29008-5715-4a39-b0ea-252402970cdc" containerID="b5c623227e0851c2a7583d3c6526727451c9485f8860571c588dfc8864724dc6" exitCode=0 Oct 01 16:18:38 crc kubenswrapper[4726]: I1001 16:18:38.383562 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b942-account-create-7x8nm" event={"ID":"82d29008-5715-4a39-b0ea-252402970cdc","Type":"ContainerDied","Data":"b5c623227e0851c2a7583d3c6526727451c9485f8860571c588dfc8864724dc6"} Oct 01 16:18:38 crc kubenswrapper[4726]: I1001 16:18:38.383611 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b942-account-create-7x8nm" event={"ID":"82d29008-5715-4a39-b0ea-252402970cdc","Type":"ContainerStarted","Data":"5dd9d30832bd1af9b339db3db0cfb39acaec0e1201154dc0b83beb39a3ee0212"} Oct 01 16:18:38 crc kubenswrapper[4726]: I1001 16:18:38.383634 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 16:18:38 crc kubenswrapper[4726]: I1001 16:18:38.383646 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 16:18:39 crc kubenswrapper[4726]: I1001 16:18:39.393016 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e65728b8-4153-40f8-ad98-c3b6abacf176","Type":"ContainerStarted","Data":"f4e03de22a32083ff1e9334dad1912052f957815fdebc7d93676db859de52631"} Oct 01 16:18:39 crc kubenswrapper[4726]: I1001 16:18:39.393063 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="ceilometer-central-agent" containerID="cri-o://eb4a6ee6fafd7da92c626c4937c8b6ee39ffa9671dad6a6aee4baf659ad73063" gracePeriod=30 Oct 01 16:18:39 crc kubenswrapper[4726]: I1001 16:18:39.393091 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="sg-core" containerID="cri-o://472776a7552756900eed556bff8d94abd8bcc8c1c2f9ee3b7e2fd75376435097" gracePeriod=30 Oct 01 16:18:39 crc kubenswrapper[4726]: I1001 16:18:39.393102 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="ceilometer-notification-agent" containerID="cri-o://09d1c5c0629432eb0d489832b2dc004f400b8a2207d2409e1e2766c0795e3f23" gracePeriod=30 Oct 01 16:18:39 crc kubenswrapper[4726]: I1001 16:18:39.393115 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="proxy-httpd" containerID="cri-o://f4e03de22a32083ff1e9334dad1912052f957815fdebc7d93676db859de52631" gracePeriod=30 Oct 01 16:18:39 crc kubenswrapper[4726]: I1001 16:18:39.393786 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 16:18:39 crc kubenswrapper[4726]: I1001 16:18:39.432707 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.395600802 podStartE2EDuration="6.432681682s" podCreationTimestamp="2025-10-01 16:18:33 +0000 UTC" firstStartedPulling="2025-10-01 16:18:34.494732443 +0000 UTC m=+1167.396285020" lastFinishedPulling="2025-10-01 16:18:38.531813323 +0000 UTC m=+1171.433365900" observedRunningTime="2025-10-01 16:18:39.428488574 +0000 UTC m=+1172.330041151" watchObservedRunningTime="2025-10-01 16:18:39.432681682 +0000 UTC m=+1172.334234279" Oct 01 16:18:39 crc kubenswrapper[4726]: I1001 16:18:39.872592 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-560d-account-create-fztkr" Oct 01 16:18:39 crc kubenswrapper[4726]: I1001 16:18:39.879347 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f3eb-account-create-6g7cm" Oct 01 16:18:39 crc kubenswrapper[4726]: I1001 16:18:39.908924 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b942-account-create-7x8nm" Oct 01 16:18:39 crc kubenswrapper[4726]: I1001 16:18:39.929841 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdj6s\" (UniqueName: \"kubernetes.io/projected/a7a05e67-296b-4ed9-b63c-30c6edc152db-kube-api-access-kdj6s\") pod \"a7a05e67-296b-4ed9-b63c-30c6edc152db\" (UID: \"a7a05e67-296b-4ed9-b63c-30c6edc152db\") " Oct 01 16:18:39 crc kubenswrapper[4726]: I1001 16:18:39.930172 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhhpn\" (UniqueName: \"kubernetes.io/projected/1c7c4328-6015-4e9a-9cd8-67c1d275928e-kube-api-access-bhhpn\") pod \"1c7c4328-6015-4e9a-9cd8-67c1d275928e\" (UID: \"1c7c4328-6015-4e9a-9cd8-67c1d275928e\") " Oct 01 16:18:39 crc kubenswrapper[4726]: I1001 16:18:39.939318 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c7c4328-6015-4e9a-9cd8-67c1d275928e-kube-api-access-bhhpn" (OuterVolumeSpecName: "kube-api-access-bhhpn") pod "1c7c4328-6015-4e9a-9cd8-67c1d275928e" (UID: "1c7c4328-6015-4e9a-9cd8-67c1d275928e"). InnerVolumeSpecName "kube-api-access-bhhpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:39 crc kubenswrapper[4726]: I1001 16:18:39.941121 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7a05e67-296b-4ed9-b63c-30c6edc152db-kube-api-access-kdj6s" (OuterVolumeSpecName: "kube-api-access-kdj6s") pod "a7a05e67-296b-4ed9-b63c-30c6edc152db" (UID: "a7a05e67-296b-4ed9-b63c-30c6edc152db"). InnerVolumeSpecName "kube-api-access-kdj6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.031909 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nxpm\" (UniqueName: \"kubernetes.io/projected/82d29008-5715-4a39-b0ea-252402970cdc-kube-api-access-2nxpm\") pod \"82d29008-5715-4a39-b0ea-252402970cdc\" (UID: \"82d29008-5715-4a39-b0ea-252402970cdc\") " Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.032400 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdj6s\" (UniqueName: \"kubernetes.io/projected/a7a05e67-296b-4ed9-b63c-30c6edc152db-kube-api-access-kdj6s\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.032417 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhhpn\" (UniqueName: \"kubernetes.io/projected/1c7c4328-6015-4e9a-9cd8-67c1d275928e-kube-api-access-bhhpn\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.035112 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82d29008-5715-4a39-b0ea-252402970cdc-kube-api-access-2nxpm" (OuterVolumeSpecName: "kube-api-access-2nxpm") pod "82d29008-5715-4a39-b0ea-252402970cdc" (UID: "82d29008-5715-4a39-b0ea-252402970cdc"). InnerVolumeSpecName "kube-api-access-2nxpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.116142 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.134384 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nxpm\" (UniqueName: \"kubernetes.io/projected/82d29008-5715-4a39-b0ea-252402970cdc-kube-api-access-2nxpm\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.235953 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-config\") pod \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.236039 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-combined-ca-bundle\") pod \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.236145 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-httpd-config\") pod \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.236187 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4rjk\" (UniqueName: \"kubernetes.io/projected/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-kube-api-access-t4rjk\") pod \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.236243 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-ovndb-tls-certs\") pod \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\" (UID: \"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984\") " Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.239746 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" (UID: "e8b0e5e1-dcbc-4118-99c1-10e1b6eea984"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.239842 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-kube-api-access-t4rjk" (OuterVolumeSpecName: "kube-api-access-t4rjk") pod "e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" (UID: "e8b0e5e1-dcbc-4118-99c1-10e1b6eea984"). InnerVolumeSpecName "kube-api-access-t4rjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.314509 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" (UID: "e8b0e5e1-dcbc-4118-99c1-10e1b6eea984"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.316928 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" (UID: "e8b0e5e1-dcbc-4118-99c1-10e1b6eea984"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.317480 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-config" (OuterVolumeSpecName: "config") pod "e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" (UID: "e8b0e5e1-dcbc-4118-99c1-10e1b6eea984"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.338982 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.339024 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.339065 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4rjk\" (UniqueName: \"kubernetes.io/projected/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-kube-api-access-t4rjk\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.339082 4726 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.339100 4726 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.410020 4726 generic.go:334] "Generic (PLEG): container finished" podID="e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" containerID="bd05040824b5af9e2be2be07f72261af0136cfbbfca823ddfc3960a35db693bf" exitCode=0 Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.410099 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-566c76dfb4-x49km" event={"ID":"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984","Type":"ContainerDied","Data":"bd05040824b5af9e2be2be07f72261af0136cfbbfca823ddfc3960a35db693bf"} Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.410128 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-566c76dfb4-x49km" event={"ID":"e8b0e5e1-dcbc-4118-99c1-10e1b6eea984","Type":"ContainerDied","Data":"de92528fc15954f4d81471dcab7ecafea712169a7ef79c4915d1d95a820c391e"} Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.410144 4726 scope.go:117] "RemoveContainer" containerID="eae9e5f7a8a4413ed91ee774a7f3ec41fe23fab3f71b4aca60e097647f504fb5" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.410542 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-566c76dfb4-x49km" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.412346 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f3eb-account-create-6g7cm" event={"ID":"a7a05e67-296b-4ed9-b63c-30c6edc152db","Type":"ContainerDied","Data":"7007d0b56c16e5086399bc0089f2a1d31a296d20cecc12708e5bb33d6821729b"} Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.412368 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7007d0b56c16e5086399bc0089f2a1d31a296d20cecc12708e5bb33d6821729b" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.412416 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f3eb-account-create-6g7cm" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.417654 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b942-account-create-7x8nm" event={"ID":"82d29008-5715-4a39-b0ea-252402970cdc","Type":"ContainerDied","Data":"5dd9d30832bd1af9b339db3db0cfb39acaec0e1201154dc0b83beb39a3ee0212"} Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.417674 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dd9d30832bd1af9b339db3db0cfb39acaec0e1201154dc0b83beb39a3ee0212" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.417709 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b942-account-create-7x8nm" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.419869 4726 generic.go:334] "Generic (PLEG): container finished" podID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerID="f4e03de22a32083ff1e9334dad1912052f957815fdebc7d93676db859de52631" exitCode=0 Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.419904 4726 generic.go:334] "Generic (PLEG): container finished" podID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerID="472776a7552756900eed556bff8d94abd8bcc8c1c2f9ee3b7e2fd75376435097" exitCode=2 Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.419916 4726 generic.go:334] "Generic (PLEG): container finished" podID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerID="09d1c5c0629432eb0d489832b2dc004f400b8a2207d2409e1e2766c0795e3f23" exitCode=0 Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.419962 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e65728b8-4153-40f8-ad98-c3b6abacf176","Type":"ContainerDied","Data":"f4e03de22a32083ff1e9334dad1912052f957815fdebc7d93676db859de52631"} Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.419989 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e65728b8-4153-40f8-ad98-c3b6abacf176","Type":"ContainerDied","Data":"472776a7552756900eed556bff8d94abd8bcc8c1c2f9ee3b7e2fd75376435097"} Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.420012 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e65728b8-4153-40f8-ad98-c3b6abacf176","Type":"ContainerDied","Data":"09d1c5c0629432eb0d489832b2dc004f400b8a2207d2409e1e2766c0795e3f23"} Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.421919 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.421942 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.423330 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-560d-account-create-fztkr" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.426360 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-560d-account-create-fztkr" event={"ID":"1c7c4328-6015-4e9a-9cd8-67c1d275928e","Type":"ContainerDied","Data":"504f67c9d9acf76055509e758242536957f01df42dc4627949c8b3a28394d992"} Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.426544 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="504f67c9d9acf76055509e758242536957f01df42dc4627949c8b3a28394d992" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.448595 4726 scope.go:117] "RemoveContainer" containerID="bd05040824b5af9e2be2be07f72261af0136cfbbfca823ddfc3960a35db693bf" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.473642 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.475840 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-566c76dfb4-x49km"] Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.481946 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.484997 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-566c76dfb4-x49km"] Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.485918 4726 scope.go:117] "RemoveContainer" containerID="eae9e5f7a8a4413ed91ee774a7f3ec41fe23fab3f71b4aca60e097647f504fb5" Oct 01 16:18:40 crc kubenswrapper[4726]: E1001 16:18:40.486427 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eae9e5f7a8a4413ed91ee774a7f3ec41fe23fab3f71b4aca60e097647f504fb5\": container with ID starting with eae9e5f7a8a4413ed91ee774a7f3ec41fe23fab3f71b4aca60e097647f504fb5 not found: ID does not exist" containerID="eae9e5f7a8a4413ed91ee774a7f3ec41fe23fab3f71b4aca60e097647f504fb5" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.486483 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eae9e5f7a8a4413ed91ee774a7f3ec41fe23fab3f71b4aca60e097647f504fb5"} err="failed to get container status \"eae9e5f7a8a4413ed91ee774a7f3ec41fe23fab3f71b4aca60e097647f504fb5\": rpc error: code = NotFound desc = could not find container \"eae9e5f7a8a4413ed91ee774a7f3ec41fe23fab3f71b4aca60e097647f504fb5\": container with ID starting with eae9e5f7a8a4413ed91ee774a7f3ec41fe23fab3f71b4aca60e097647f504fb5 not found: ID does not exist" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.486517 4726 scope.go:117] "RemoveContainer" containerID="bd05040824b5af9e2be2be07f72261af0136cfbbfca823ddfc3960a35db693bf" Oct 01 16:18:40 crc kubenswrapper[4726]: E1001 16:18:40.486795 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd05040824b5af9e2be2be07f72261af0136cfbbfca823ddfc3960a35db693bf\": container with ID starting with bd05040824b5af9e2be2be07f72261af0136cfbbfca823ddfc3960a35db693bf not found: ID does not exist" containerID="bd05040824b5af9e2be2be07f72261af0136cfbbfca823ddfc3960a35db693bf" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.486830 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd05040824b5af9e2be2be07f72261af0136cfbbfca823ddfc3960a35db693bf"} err="failed to get container status \"bd05040824b5af9e2be2be07f72261af0136cfbbfca823ddfc3960a35db693bf\": rpc error: code = NotFound desc = could not find container \"bd05040824b5af9e2be2be07f72261af0136cfbbfca823ddfc3960a35db693bf\": container with ID starting with bd05040824b5af9e2be2be07f72261af0136cfbbfca823ddfc3960a35db693bf not found: ID does not exist" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.689936 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.689991 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.747588 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 16:18:40 crc kubenswrapper[4726]: I1001 16:18:40.749622 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.435835 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.436257 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.822821 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" path="/var/lib/kubelet/pods/e8b0e5e1-dcbc-4118-99c1-10e1b6eea984/volumes" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.909181 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hq8t6"] Oct 01 16:18:41 crc kubenswrapper[4726]: E1001 16:18:41.911785 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c7c4328-6015-4e9a-9cd8-67c1d275928e" containerName="mariadb-account-create" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.911821 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c7c4328-6015-4e9a-9cd8-67c1d275928e" containerName="mariadb-account-create" Oct 01 16:18:41 crc kubenswrapper[4726]: E1001 16:18:41.911834 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" containerName="neutron-httpd" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.911842 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" containerName="neutron-httpd" Oct 01 16:18:41 crc kubenswrapper[4726]: E1001 16:18:41.911869 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82d29008-5715-4a39-b0ea-252402970cdc" containerName="mariadb-account-create" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.911878 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="82d29008-5715-4a39-b0ea-252402970cdc" containerName="mariadb-account-create" Oct 01 16:18:41 crc kubenswrapper[4726]: E1001 16:18:41.911917 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7a05e67-296b-4ed9-b63c-30c6edc152db" containerName="mariadb-account-create" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.911926 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7a05e67-296b-4ed9-b63c-30c6edc152db" containerName="mariadb-account-create" Oct 01 16:18:41 crc kubenswrapper[4726]: E1001 16:18:41.911949 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" containerName="neutron-api" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.911957 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" containerName="neutron-api" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.912201 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c7c4328-6015-4e9a-9cd8-67c1d275928e" containerName="mariadb-account-create" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.912228 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" containerName="neutron-httpd" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.912241 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8b0e5e1-dcbc-4118-99c1-10e1b6eea984" containerName="neutron-api" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.912259 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="82d29008-5715-4a39-b0ea-252402970cdc" containerName="mariadb-account-create" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.912272 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7a05e67-296b-4ed9-b63c-30c6edc152db" containerName="mariadb-account-create" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.913000 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.915989 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.920546 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.920741 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-pdwp7" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.948756 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hq8t6"] Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.982962 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hq8t6\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.983077 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-config-data\") pod \"nova-cell0-conductor-db-sync-hq8t6\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.983235 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-scripts\") pod \"nova-cell0-conductor-db-sync-hq8t6\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:18:41 crc kubenswrapper[4726]: I1001 16:18:41.983264 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75v2t\" (UniqueName: \"kubernetes.io/projected/551173b5-0f09-4b1e-831b-c6598ba904cb-kube-api-access-75v2t\") pod \"nova-cell0-conductor-db-sync-hq8t6\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.092412 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hq8t6\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.092521 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-config-data\") pod \"nova-cell0-conductor-db-sync-hq8t6\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.092667 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-scripts\") pod \"nova-cell0-conductor-db-sync-hq8t6\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.092718 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75v2t\" (UniqueName: \"kubernetes.io/projected/551173b5-0f09-4b1e-831b-c6598ba904cb-kube-api-access-75v2t\") pod \"nova-cell0-conductor-db-sync-hq8t6\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.101664 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-scripts\") pod \"nova-cell0-conductor-db-sync-hq8t6\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.116730 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-config-data\") pod \"nova-cell0-conductor-db-sync-hq8t6\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.129733 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hq8t6\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.163645 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75v2t\" (UniqueName: \"kubernetes.io/projected/551173b5-0f09-4b1e-831b-c6598ba904cb-kube-api-access-75v2t\") pod \"nova-cell0-conductor-db-sync-hq8t6\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.256823 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.294287 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.398403 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e65728b8-4153-40f8-ad98-c3b6abacf176-run-httpd\") pod \"e65728b8-4153-40f8-ad98-c3b6abacf176\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.399027 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-scripts\") pod \"e65728b8-4153-40f8-ad98-c3b6abacf176\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.399786 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-config-data\") pod \"e65728b8-4153-40f8-ad98-c3b6abacf176\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.398976 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e65728b8-4153-40f8-ad98-c3b6abacf176-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e65728b8-4153-40f8-ad98-c3b6abacf176" (UID: "e65728b8-4153-40f8-ad98-c3b6abacf176"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.399968 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qjds\" (UniqueName: \"kubernetes.io/projected/e65728b8-4153-40f8-ad98-c3b6abacf176-kube-api-access-7qjds\") pod \"e65728b8-4153-40f8-ad98-c3b6abacf176\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.400008 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e65728b8-4153-40f8-ad98-c3b6abacf176-log-httpd\") pod \"e65728b8-4153-40f8-ad98-c3b6abacf176\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.400092 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-sg-core-conf-yaml\") pod \"e65728b8-4153-40f8-ad98-c3b6abacf176\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.400120 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-combined-ca-bundle\") pod \"e65728b8-4153-40f8-ad98-c3b6abacf176\" (UID: \"e65728b8-4153-40f8-ad98-c3b6abacf176\") " Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.400463 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e65728b8-4153-40f8-ad98-c3b6abacf176-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e65728b8-4153-40f8-ad98-c3b6abacf176" (UID: "e65728b8-4153-40f8-ad98-c3b6abacf176"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.401102 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e65728b8-4153-40f8-ad98-c3b6abacf176-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.401125 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e65728b8-4153-40f8-ad98-c3b6abacf176-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.405405 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-scripts" (OuterVolumeSpecName: "scripts") pod "e65728b8-4153-40f8-ad98-c3b6abacf176" (UID: "e65728b8-4153-40f8-ad98-c3b6abacf176"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.405541 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e65728b8-4153-40f8-ad98-c3b6abacf176-kube-api-access-7qjds" (OuterVolumeSpecName: "kube-api-access-7qjds") pod "e65728b8-4153-40f8-ad98-c3b6abacf176" (UID: "e65728b8-4153-40f8-ad98-c3b6abacf176"). InnerVolumeSpecName "kube-api-access-7qjds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.449626 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e65728b8-4153-40f8-ad98-c3b6abacf176" (UID: "e65728b8-4153-40f8-ad98-c3b6abacf176"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.468340 4726 generic.go:334] "Generic (PLEG): container finished" podID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerID="eb4a6ee6fafd7da92c626c4937c8b6ee39ffa9671dad6a6aee4baf659ad73063" exitCode=0 Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.469392 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.469930 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e65728b8-4153-40f8-ad98-c3b6abacf176","Type":"ContainerDied","Data":"eb4a6ee6fafd7da92c626c4937c8b6ee39ffa9671dad6a6aee4baf659ad73063"} Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.469967 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e65728b8-4153-40f8-ad98-c3b6abacf176","Type":"ContainerDied","Data":"65b82d1fee7e578b3617d9ecf87e555563acb46c55ca394c0cd5d0f657ba299c"} Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.469986 4726 scope.go:117] "RemoveContainer" containerID="f4e03de22a32083ff1e9334dad1912052f957815fdebc7d93676db859de52631" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.502221 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.502456 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.502465 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qjds\" (UniqueName: \"kubernetes.io/projected/e65728b8-4153-40f8-ad98-c3b6abacf176-kube-api-access-7qjds\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.514233 4726 scope.go:117] "RemoveContainer" containerID="472776a7552756900eed556bff8d94abd8bcc8c1c2f9ee3b7e2fd75376435097" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.534351 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-config-data" (OuterVolumeSpecName: "config-data") pod "e65728b8-4153-40f8-ad98-c3b6abacf176" (UID: "e65728b8-4153-40f8-ad98-c3b6abacf176"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.537041 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e65728b8-4153-40f8-ad98-c3b6abacf176" (UID: "e65728b8-4153-40f8-ad98-c3b6abacf176"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.550774 4726 scope.go:117] "RemoveContainer" containerID="09d1c5c0629432eb0d489832b2dc004f400b8a2207d2409e1e2766c0795e3f23" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.579542 4726 scope.go:117] "RemoveContainer" containerID="eb4a6ee6fafd7da92c626c4937c8b6ee39ffa9671dad6a6aee4baf659ad73063" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.604516 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.604605 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e65728b8-4153-40f8-ad98-c3b6abacf176-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.607704 4726 scope.go:117] "RemoveContainer" containerID="f4e03de22a32083ff1e9334dad1912052f957815fdebc7d93676db859de52631" Oct 01 16:18:42 crc kubenswrapper[4726]: E1001 16:18:42.608470 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4e03de22a32083ff1e9334dad1912052f957815fdebc7d93676db859de52631\": container with ID starting with f4e03de22a32083ff1e9334dad1912052f957815fdebc7d93676db859de52631 not found: ID does not exist" containerID="f4e03de22a32083ff1e9334dad1912052f957815fdebc7d93676db859de52631" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.608524 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4e03de22a32083ff1e9334dad1912052f957815fdebc7d93676db859de52631"} err="failed to get container status \"f4e03de22a32083ff1e9334dad1912052f957815fdebc7d93676db859de52631\": rpc error: code = NotFound desc = could not find container \"f4e03de22a32083ff1e9334dad1912052f957815fdebc7d93676db859de52631\": container with ID starting with f4e03de22a32083ff1e9334dad1912052f957815fdebc7d93676db859de52631 not found: ID does not exist" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.608555 4726 scope.go:117] "RemoveContainer" containerID="472776a7552756900eed556bff8d94abd8bcc8c1c2f9ee3b7e2fd75376435097" Oct 01 16:18:42 crc kubenswrapper[4726]: E1001 16:18:42.608962 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"472776a7552756900eed556bff8d94abd8bcc8c1c2f9ee3b7e2fd75376435097\": container with ID starting with 472776a7552756900eed556bff8d94abd8bcc8c1c2f9ee3b7e2fd75376435097 not found: ID does not exist" containerID="472776a7552756900eed556bff8d94abd8bcc8c1c2f9ee3b7e2fd75376435097" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.608992 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"472776a7552756900eed556bff8d94abd8bcc8c1c2f9ee3b7e2fd75376435097"} err="failed to get container status \"472776a7552756900eed556bff8d94abd8bcc8c1c2f9ee3b7e2fd75376435097\": rpc error: code = NotFound desc = could not find container \"472776a7552756900eed556bff8d94abd8bcc8c1c2f9ee3b7e2fd75376435097\": container with ID starting with 472776a7552756900eed556bff8d94abd8bcc8c1c2f9ee3b7e2fd75376435097 not found: ID does not exist" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.609029 4726 scope.go:117] "RemoveContainer" containerID="09d1c5c0629432eb0d489832b2dc004f400b8a2207d2409e1e2766c0795e3f23" Oct 01 16:18:42 crc kubenswrapper[4726]: E1001 16:18:42.609361 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09d1c5c0629432eb0d489832b2dc004f400b8a2207d2409e1e2766c0795e3f23\": container with ID starting with 09d1c5c0629432eb0d489832b2dc004f400b8a2207d2409e1e2766c0795e3f23 not found: ID does not exist" containerID="09d1c5c0629432eb0d489832b2dc004f400b8a2207d2409e1e2766c0795e3f23" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.609390 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09d1c5c0629432eb0d489832b2dc004f400b8a2207d2409e1e2766c0795e3f23"} err="failed to get container status \"09d1c5c0629432eb0d489832b2dc004f400b8a2207d2409e1e2766c0795e3f23\": rpc error: code = NotFound desc = could not find container \"09d1c5c0629432eb0d489832b2dc004f400b8a2207d2409e1e2766c0795e3f23\": container with ID starting with 09d1c5c0629432eb0d489832b2dc004f400b8a2207d2409e1e2766c0795e3f23 not found: ID does not exist" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.609407 4726 scope.go:117] "RemoveContainer" containerID="eb4a6ee6fafd7da92c626c4937c8b6ee39ffa9671dad6a6aee4baf659ad73063" Oct 01 16:18:42 crc kubenswrapper[4726]: E1001 16:18:42.609813 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb4a6ee6fafd7da92c626c4937c8b6ee39ffa9671dad6a6aee4baf659ad73063\": container with ID starting with eb4a6ee6fafd7da92c626c4937c8b6ee39ffa9671dad6a6aee4baf659ad73063 not found: ID does not exist" containerID="eb4a6ee6fafd7da92c626c4937c8b6ee39ffa9671dad6a6aee4baf659ad73063" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.609839 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb4a6ee6fafd7da92c626c4937c8b6ee39ffa9671dad6a6aee4baf659ad73063"} err="failed to get container status \"eb4a6ee6fafd7da92c626c4937c8b6ee39ffa9671dad6a6aee4baf659ad73063\": rpc error: code = NotFound desc = could not find container \"eb4a6ee6fafd7da92c626c4937c8b6ee39ffa9671dad6a6aee4baf659ad73063\": container with ID starting with eb4a6ee6fafd7da92c626c4937c8b6ee39ffa9671dad6a6aee4baf659ad73063 not found: ID does not exist" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.737682 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hq8t6"] Oct 01 16:18:42 crc kubenswrapper[4726]: W1001 16:18:42.738469 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod551173b5_0f09_4b1e_831b_c6598ba904cb.slice/crio-b7bb5e0b34a81f9494c7c74f8588c00e3036e68b71eee3d196d036d2e84d0c31 WatchSource:0}: Error finding container b7bb5e0b34a81f9494c7c74f8588c00e3036e68b71eee3d196d036d2e84d0c31: Status 404 returned error can't find the container with id b7bb5e0b34a81f9494c7c74f8588c00e3036e68b71eee3d196d036d2e84d0c31 Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.825330 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.833163 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.867846 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:42 crc kubenswrapper[4726]: E1001 16:18:42.868315 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="proxy-httpd" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.868337 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="proxy-httpd" Oct 01 16:18:42 crc kubenswrapper[4726]: E1001 16:18:42.868360 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="ceilometer-central-agent" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.868369 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="ceilometer-central-agent" Oct 01 16:18:42 crc kubenswrapper[4726]: E1001 16:18:42.868410 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="ceilometer-notification-agent" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.868419 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="ceilometer-notification-agent" Oct 01 16:18:42 crc kubenswrapper[4726]: E1001 16:18:42.868432 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="sg-core" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.868441 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="sg-core" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.868678 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="ceilometer-notification-agent" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.868697 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="ceilometer-central-agent" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.868719 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="sg-core" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.868735 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" containerName="proxy-httpd" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.871270 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.886785 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.892546 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.892923 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.908381 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3da892a-9963-4863-a9b7-5d50ead044cd-run-httpd\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.908789 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3da892a-9963-4863-a9b7-5d50ead044cd-log-httpd\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.908846 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-config-data\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.909016 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.909123 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwzm8\" (UniqueName: \"kubernetes.io/projected/b3da892a-9963-4863-a9b7-5d50ead044cd-kube-api-access-cwzm8\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.909164 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-scripts\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:42 crc kubenswrapper[4726]: I1001 16:18:42.909211 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.010730 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.011569 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3da892a-9963-4863-a9b7-5d50ead044cd-run-httpd\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.011611 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3da892a-9963-4863-a9b7-5d50ead044cd-log-httpd\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.011637 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-config-data\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.011727 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.011771 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwzm8\" (UniqueName: \"kubernetes.io/projected/b3da892a-9963-4863-a9b7-5d50ead044cd-kube-api-access-cwzm8\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.011788 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-scripts\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.012559 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3da892a-9963-4863-a9b7-5d50ead044cd-log-httpd\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.012570 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3da892a-9963-4863-a9b7-5d50ead044cd-run-httpd\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.017508 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-config-data\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.019532 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.033565 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-scripts\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.033566 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwzm8\" (UniqueName: \"kubernetes.io/projected/b3da892a-9963-4863-a9b7-5d50ead044cd-kube-api-access-cwzm8\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.033733 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.213999 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.525469 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hq8t6" event={"ID":"551173b5-0f09-4b1e-831b-c6598ba904cb","Type":"ContainerStarted","Data":"b7bb5e0b34a81f9494c7c74f8588c00e3036e68b71eee3d196d036d2e84d0c31"} Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.527422 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.527438 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.698175 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:43 crc kubenswrapper[4726]: W1001 16:18:43.720498 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3da892a_9963_4863_a9b7_5d50ead044cd.slice/crio-e912bd18010fd4eea7d47b7b373b6e8921dfd1e1155709de36a801a05bf08a05 WatchSource:0}: Error finding container e912bd18010fd4eea7d47b7b373b6e8921dfd1e1155709de36a801a05bf08a05: Status 404 returned error can't find the container with id e912bd18010fd4eea7d47b7b373b6e8921dfd1e1155709de36a801a05bf08a05 Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.801513 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.829653 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e65728b8-4153-40f8-ad98-c3b6abacf176" path="/var/lib/kubelet/pods/e65728b8-4153-40f8-ad98-c3b6abacf176/volumes" Oct 01 16:18:43 crc kubenswrapper[4726]: I1001 16:18:43.926133 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 16:18:44 crc kubenswrapper[4726]: I1001 16:18:44.558643 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3da892a-9963-4863-a9b7-5d50ead044cd","Type":"ContainerStarted","Data":"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b"} Oct 01 16:18:44 crc kubenswrapper[4726]: I1001 16:18:44.558689 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3da892a-9963-4863-a9b7-5d50ead044cd","Type":"ContainerStarted","Data":"e912bd18010fd4eea7d47b7b373b6e8921dfd1e1155709de36a801a05bf08a05"} Oct 01 16:18:44 crc kubenswrapper[4726]: I1001 16:18:44.937389 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:45 crc kubenswrapper[4726]: I1001 16:18:45.581301 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3da892a-9963-4863-a9b7-5d50ead044cd","Type":"ContainerStarted","Data":"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056"} Oct 01 16:18:46 crc kubenswrapper[4726]: I1001 16:18:46.601319 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3da892a-9963-4863-a9b7-5d50ead044cd","Type":"ContainerStarted","Data":"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe"} Oct 01 16:18:51 crc kubenswrapper[4726]: I1001 16:18:51.651594 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3da892a-9963-4863-a9b7-5d50ead044cd","Type":"ContainerStarted","Data":"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f"} Oct 01 16:18:51 crc kubenswrapper[4726]: I1001 16:18:51.651901 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="ceilometer-central-agent" containerID="cri-o://25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b" gracePeriod=30 Oct 01 16:18:51 crc kubenswrapper[4726]: I1001 16:18:51.652412 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 16:18:51 crc kubenswrapper[4726]: I1001 16:18:51.653721 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="proxy-httpd" containerID="cri-o://00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f" gracePeriod=30 Oct 01 16:18:51 crc kubenswrapper[4726]: I1001 16:18:51.653976 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="sg-core" containerID="cri-o://320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe" gracePeriod=30 Oct 01 16:18:51 crc kubenswrapper[4726]: I1001 16:18:51.655310 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="ceilometer-notification-agent" containerID="cri-o://63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056" gracePeriod=30 Oct 01 16:18:51 crc kubenswrapper[4726]: I1001 16:18:51.665372 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hq8t6" event={"ID":"551173b5-0f09-4b1e-831b-c6598ba904cb","Type":"ContainerStarted","Data":"7dbe378442bb110761661bc29105a34b7e542487981a30e0ee6ff6c42928e6cd"} Oct 01 16:18:51 crc kubenswrapper[4726]: I1001 16:18:51.700680 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.076532302 podStartE2EDuration="9.700650121s" podCreationTimestamp="2025-10-01 16:18:42 +0000 UTC" firstStartedPulling="2025-10-01 16:18:43.725704962 +0000 UTC m=+1176.627257539" lastFinishedPulling="2025-10-01 16:18:50.349822771 +0000 UTC m=+1183.251375358" observedRunningTime="2025-10-01 16:18:51.692545112 +0000 UTC m=+1184.594097699" watchObservedRunningTime="2025-10-01 16:18:51.700650121 +0000 UTC m=+1184.602202738" Oct 01 16:18:51 crc kubenswrapper[4726]: I1001 16:18:51.726099 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-hq8t6" podStartSLOduration=3.10887304 podStartE2EDuration="10.726074877s" podCreationTimestamp="2025-10-01 16:18:41 +0000 UTC" firstStartedPulling="2025-10-01 16:18:42.740327301 +0000 UTC m=+1175.641879888" lastFinishedPulling="2025-10-01 16:18:50.357529148 +0000 UTC m=+1183.259081725" observedRunningTime="2025-10-01 16:18:51.716407545 +0000 UTC m=+1184.617960192" watchObservedRunningTime="2025-10-01 16:18:51.726074877 +0000 UTC m=+1184.627627464" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.397223 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.508741 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3da892a-9963-4863-a9b7-5d50ead044cd-run-httpd\") pod \"b3da892a-9963-4863-a9b7-5d50ead044cd\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.508787 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-sg-core-conf-yaml\") pod \"b3da892a-9963-4863-a9b7-5d50ead044cd\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.508870 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3da892a-9963-4863-a9b7-5d50ead044cd-log-httpd\") pod \"b3da892a-9963-4863-a9b7-5d50ead044cd\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.509311 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3da892a-9963-4863-a9b7-5d50ead044cd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b3da892a-9963-4863-a9b7-5d50ead044cd" (UID: "b3da892a-9963-4863-a9b7-5d50ead044cd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.509420 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3da892a-9963-4863-a9b7-5d50ead044cd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b3da892a-9963-4863-a9b7-5d50ead044cd" (UID: "b3da892a-9963-4863-a9b7-5d50ead044cd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.509503 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-combined-ca-bundle\") pod \"b3da892a-9963-4863-a9b7-5d50ead044cd\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.509544 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-scripts\") pod \"b3da892a-9963-4863-a9b7-5d50ead044cd\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.509693 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwzm8\" (UniqueName: \"kubernetes.io/projected/b3da892a-9963-4863-a9b7-5d50ead044cd-kube-api-access-cwzm8\") pod \"b3da892a-9963-4863-a9b7-5d50ead044cd\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.509855 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-config-data\") pod \"b3da892a-9963-4863-a9b7-5d50ead044cd\" (UID: \"b3da892a-9963-4863-a9b7-5d50ead044cd\") " Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.510356 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3da892a-9963-4863-a9b7-5d50ead044cd-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.510378 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3da892a-9963-4863-a9b7-5d50ead044cd-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.515107 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3da892a-9963-4863-a9b7-5d50ead044cd-kube-api-access-cwzm8" (OuterVolumeSpecName: "kube-api-access-cwzm8") pod "b3da892a-9963-4863-a9b7-5d50ead044cd" (UID: "b3da892a-9963-4863-a9b7-5d50ead044cd"). InnerVolumeSpecName "kube-api-access-cwzm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.515295 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-scripts" (OuterVolumeSpecName: "scripts") pod "b3da892a-9963-4863-a9b7-5d50ead044cd" (UID: "b3da892a-9963-4863-a9b7-5d50ead044cd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.540754 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b3da892a-9963-4863-a9b7-5d50ead044cd" (UID: "b3da892a-9963-4863-a9b7-5d50ead044cd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.579763 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3da892a-9963-4863-a9b7-5d50ead044cd" (UID: "b3da892a-9963-4863-a9b7-5d50ead044cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.604351 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-config-data" (OuterVolumeSpecName: "config-data") pod "b3da892a-9963-4863-a9b7-5d50ead044cd" (UID: "b3da892a-9963-4863-a9b7-5d50ead044cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.612198 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwzm8\" (UniqueName: \"kubernetes.io/projected/b3da892a-9963-4863-a9b7-5d50ead044cd-kube-api-access-cwzm8\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.612245 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.612258 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.612270 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.612281 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3da892a-9963-4863-a9b7-5d50ead044cd-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.677077 4726 generic.go:334] "Generic (PLEG): container finished" podID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerID="00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f" exitCode=0 Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.677107 4726 generic.go:334] "Generic (PLEG): container finished" podID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerID="320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe" exitCode=2 Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.677114 4726 generic.go:334] "Generic (PLEG): container finished" podID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerID="63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056" exitCode=0 Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.677122 4726 generic.go:334] "Generic (PLEG): container finished" podID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerID="25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b" exitCode=0 Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.677990 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.679286 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3da892a-9963-4863-a9b7-5d50ead044cd","Type":"ContainerDied","Data":"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f"} Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.679509 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3da892a-9963-4863-a9b7-5d50ead044cd","Type":"ContainerDied","Data":"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe"} Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.679525 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3da892a-9963-4863-a9b7-5d50ead044cd","Type":"ContainerDied","Data":"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056"} Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.679543 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3da892a-9963-4863-a9b7-5d50ead044cd","Type":"ContainerDied","Data":"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b"} Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.679555 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3da892a-9963-4863-a9b7-5d50ead044cd","Type":"ContainerDied","Data":"e912bd18010fd4eea7d47b7b373b6e8921dfd1e1155709de36a801a05bf08a05"} Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.679602 4726 scope.go:117] "RemoveContainer" containerID="00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.714168 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.721839 4726 scope.go:117] "RemoveContainer" containerID="320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.734853 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.750302 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:52 crc kubenswrapper[4726]: E1001 16:18:52.750886 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="ceilometer-notification-agent" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.750915 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="ceilometer-notification-agent" Oct 01 16:18:52 crc kubenswrapper[4726]: E1001 16:18:52.750944 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="sg-core" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.750953 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="sg-core" Oct 01 16:18:52 crc kubenswrapper[4726]: E1001 16:18:52.750974 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="proxy-httpd" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.750982 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="proxy-httpd" Oct 01 16:18:52 crc kubenswrapper[4726]: E1001 16:18:52.751000 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="ceilometer-central-agent" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.751009 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="ceilometer-central-agent" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.751298 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="proxy-httpd" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.751332 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="sg-core" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.751352 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="ceilometer-notification-agent" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.751378 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" containerName="ceilometer-central-agent" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.752875 4726 scope.go:117] "RemoveContainer" containerID="63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.753820 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.758037 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.758305 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.761061 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.816798 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-scripts\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.816857 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.816886 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82fecdc7-09c8-4f55-aa63-5e953ada8783-run-httpd\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.817014 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.817074 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-config-data\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.817426 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrtj8\" (UniqueName: \"kubernetes.io/projected/82fecdc7-09c8-4f55-aa63-5e953ada8783-kube-api-access-rrtj8\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.817462 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82fecdc7-09c8-4f55-aa63-5e953ada8783-log-httpd\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.841732 4726 scope.go:117] "RemoveContainer" containerID="25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.861320 4726 scope.go:117] "RemoveContainer" containerID="00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f" Oct 01 16:18:52 crc kubenswrapper[4726]: E1001 16:18:52.861674 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f\": container with ID starting with 00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f not found: ID does not exist" containerID="00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.861712 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f"} err="failed to get container status \"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f\": rpc error: code = NotFound desc = could not find container \"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f\": container with ID starting with 00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.861737 4726 scope.go:117] "RemoveContainer" containerID="320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe" Oct 01 16:18:52 crc kubenswrapper[4726]: E1001 16:18:52.861941 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe\": container with ID starting with 320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe not found: ID does not exist" containerID="320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.861956 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe"} err="failed to get container status \"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe\": rpc error: code = NotFound desc = could not find container \"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe\": container with ID starting with 320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.861971 4726 scope.go:117] "RemoveContainer" containerID="63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056" Oct 01 16:18:52 crc kubenswrapper[4726]: E1001 16:18:52.862177 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056\": container with ID starting with 63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056 not found: ID does not exist" containerID="63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.862199 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056"} err="failed to get container status \"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056\": rpc error: code = NotFound desc = could not find container \"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056\": container with ID starting with 63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056 not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.862211 4726 scope.go:117] "RemoveContainer" containerID="25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b" Oct 01 16:18:52 crc kubenswrapper[4726]: E1001 16:18:52.862387 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b\": container with ID starting with 25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b not found: ID does not exist" containerID="25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.862403 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b"} err="failed to get container status \"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b\": rpc error: code = NotFound desc = could not find container \"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b\": container with ID starting with 25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.862416 4726 scope.go:117] "RemoveContainer" containerID="00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.862581 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f"} err="failed to get container status \"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f\": rpc error: code = NotFound desc = could not find container \"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f\": container with ID starting with 00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.862594 4726 scope.go:117] "RemoveContainer" containerID="320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.862762 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe"} err="failed to get container status \"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe\": rpc error: code = NotFound desc = could not find container \"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe\": container with ID starting with 320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.862779 4726 scope.go:117] "RemoveContainer" containerID="63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.862939 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056"} err="failed to get container status \"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056\": rpc error: code = NotFound desc = could not find container \"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056\": container with ID starting with 63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056 not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.862953 4726 scope.go:117] "RemoveContainer" containerID="25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.863154 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b"} err="failed to get container status \"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b\": rpc error: code = NotFound desc = could not find container \"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b\": container with ID starting with 25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.863171 4726 scope.go:117] "RemoveContainer" containerID="00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.863343 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f"} err="failed to get container status \"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f\": rpc error: code = NotFound desc = could not find container \"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f\": container with ID starting with 00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.863357 4726 scope.go:117] "RemoveContainer" containerID="320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.863522 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe"} err="failed to get container status \"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe\": rpc error: code = NotFound desc = could not find container \"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe\": container with ID starting with 320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.863536 4726 scope.go:117] "RemoveContainer" containerID="63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.863837 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056"} err="failed to get container status \"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056\": rpc error: code = NotFound desc = could not find container \"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056\": container with ID starting with 63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056 not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.863854 4726 scope.go:117] "RemoveContainer" containerID="25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.864030 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b"} err="failed to get container status \"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b\": rpc error: code = NotFound desc = could not find container \"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b\": container with ID starting with 25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.864044 4726 scope.go:117] "RemoveContainer" containerID="00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.864782 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f"} err="failed to get container status \"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f\": rpc error: code = NotFound desc = could not find container \"00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f\": container with ID starting with 00f5e8939b992f3088aaf504da4dd3d068d32bc64c01c8fcbbab3bbf1c9fea3f not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.864807 4726 scope.go:117] "RemoveContainer" containerID="320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.865019 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe"} err="failed to get container status \"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe\": rpc error: code = NotFound desc = could not find container \"320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe\": container with ID starting with 320e2e0bc3ca74735f76fbeab71c0866b998d6b0fe003dee5a111831b9c2cbfe not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.865039 4726 scope.go:117] "RemoveContainer" containerID="63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.865247 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056"} err="failed to get container status \"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056\": rpc error: code = NotFound desc = could not find container \"63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056\": container with ID starting with 63232c0d52f58f229e7b5c0c5e96c0c91daaddaa548acfc2126abe1f43653056 not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.865271 4726 scope.go:117] "RemoveContainer" containerID="25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.865468 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b"} err="failed to get container status \"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b\": rpc error: code = NotFound desc = could not find container \"25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b\": container with ID starting with 25b427a557b04a2eb9e07f17929c6f5458b1c3a9d577b06776bec46f60c6da8b not found: ID does not exist" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.918613 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-scripts\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.918668 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.918703 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82fecdc7-09c8-4f55-aa63-5e953ada8783-run-httpd\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.919155 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.919209 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-config-data\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.919317 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82fecdc7-09c8-4f55-aa63-5e953ada8783-log-httpd\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.919341 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrtj8\" (UniqueName: \"kubernetes.io/projected/82fecdc7-09c8-4f55-aa63-5e953ada8783-kube-api-access-rrtj8\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.919343 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82fecdc7-09c8-4f55-aa63-5e953ada8783-run-httpd\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.919745 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82fecdc7-09c8-4f55-aa63-5e953ada8783-log-httpd\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.924755 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.925030 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-scripts\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.925993 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-config-data\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.926531 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:52 crc kubenswrapper[4726]: I1001 16:18:52.942175 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrtj8\" (UniqueName: \"kubernetes.io/projected/82fecdc7-09c8-4f55-aa63-5e953ada8783-kube-api-access-rrtj8\") pod \"ceilometer-0\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " pod="openstack/ceilometer-0" Oct 01 16:18:53 crc kubenswrapper[4726]: I1001 16:18:53.141839 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:18:53 crc kubenswrapper[4726]: I1001 16:18:53.413147 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:18:53 crc kubenswrapper[4726]: I1001 16:18:53.413456 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:18:53 crc kubenswrapper[4726]: I1001 16:18:53.413500 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 16:18:53 crc kubenswrapper[4726]: I1001 16:18:53.414126 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"db368996f7ef78489a378108debc56732f3e0a06eb79040d4b7e667cfd8503a8"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:18:53 crc kubenswrapper[4726]: I1001 16:18:53.414181 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://db368996f7ef78489a378108debc56732f3e0a06eb79040d4b7e667cfd8503a8" gracePeriod=600 Oct 01 16:18:53 crc kubenswrapper[4726]: I1001 16:18:53.606262 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:18:53 crc kubenswrapper[4726]: W1001 16:18:53.609634 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82fecdc7_09c8_4f55_aa63_5e953ada8783.slice/crio-3906fa24eb4b6beb55bed80efb86ecba4be6d4c19214d1b19a98e9eee013cb63 WatchSource:0}: Error finding container 3906fa24eb4b6beb55bed80efb86ecba4be6d4c19214d1b19a98e9eee013cb63: Status 404 returned error can't find the container with id 3906fa24eb4b6beb55bed80efb86ecba4be6d4c19214d1b19a98e9eee013cb63 Oct 01 16:18:53 crc kubenswrapper[4726]: I1001 16:18:53.699804 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="db368996f7ef78489a378108debc56732f3e0a06eb79040d4b7e667cfd8503a8" exitCode=0 Oct 01 16:18:53 crc kubenswrapper[4726]: I1001 16:18:53.700350 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"db368996f7ef78489a378108debc56732f3e0a06eb79040d4b7e667cfd8503a8"} Oct 01 16:18:53 crc kubenswrapper[4726]: I1001 16:18:53.700414 4726 scope.go:117] "RemoveContainer" containerID="3fcc119df2ac4146cc2bd5a3dfb304e41077427fd468db2dfae53b91e3c841a2" Oct 01 16:18:53 crc kubenswrapper[4726]: I1001 16:18:53.701544 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82fecdc7-09c8-4f55-aa63-5e953ada8783","Type":"ContainerStarted","Data":"3906fa24eb4b6beb55bed80efb86ecba4be6d4c19214d1b19a98e9eee013cb63"} Oct 01 16:18:53 crc kubenswrapper[4726]: I1001 16:18:53.822987 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3da892a-9963-4863-a9b7-5d50ead044cd" path="/var/lib/kubelet/pods/b3da892a-9963-4863-a9b7-5d50ead044cd/volumes" Oct 01 16:18:54 crc kubenswrapper[4726]: I1001 16:18:54.723528 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"f2fff8833153004baace8bca4e9e33e62ef29d5bce54ddbea47fab14e44b1073"} Oct 01 16:18:55 crc kubenswrapper[4726]: I1001 16:18:55.742075 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82fecdc7-09c8-4f55-aa63-5e953ada8783","Type":"ContainerStarted","Data":"bb247c2c0d6727ac92bbefbe0691be261f8f09ea1f63430728df3189110da6a8"} Oct 01 16:18:56 crc kubenswrapper[4726]: I1001 16:18:56.755817 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82fecdc7-09c8-4f55-aa63-5e953ada8783","Type":"ContainerStarted","Data":"9e1d97f3da2ba2656715f91fa9077d540f2a066473452e65d17d17f4bb43094d"} Oct 01 16:18:57 crc kubenswrapper[4726]: I1001 16:18:57.771104 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82fecdc7-09c8-4f55-aa63-5e953ada8783","Type":"ContainerStarted","Data":"b8e75a9ae62ebfe1b87a2db2d77ba2ee76f8485f8245d5668e777dee5acb3856"} Oct 01 16:19:08 crc kubenswrapper[4726]: I1001 16:19:08.886014 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82fecdc7-09c8-4f55-aa63-5e953ada8783","Type":"ContainerStarted","Data":"7c81d977f6de6cb974b9cf597dac3ff42f98c096b8992b13aaaf112c762e5487"} Oct 01 16:19:08 crc kubenswrapper[4726]: I1001 16:19:08.886682 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 16:19:08 crc kubenswrapper[4726]: I1001 16:19:08.908192 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.5064012460000002 podStartE2EDuration="16.908173201s" podCreationTimestamp="2025-10-01 16:18:52 +0000 UTC" firstStartedPulling="2025-10-01 16:18:53.612101702 +0000 UTC m=+1186.513654279" lastFinishedPulling="2025-10-01 16:19:08.013873647 +0000 UTC m=+1200.915426234" observedRunningTime="2025-10-01 16:19:08.905637239 +0000 UTC m=+1201.807189826" watchObservedRunningTime="2025-10-01 16:19:08.908173201 +0000 UTC m=+1201.809725778" Oct 01 16:19:16 crc kubenswrapper[4726]: I1001 16:19:16.976798 4726 generic.go:334] "Generic (PLEG): container finished" podID="551173b5-0f09-4b1e-831b-c6598ba904cb" containerID="7dbe378442bb110761661bc29105a34b7e542487981a30e0ee6ff6c42928e6cd" exitCode=0 Oct 01 16:19:16 crc kubenswrapper[4726]: I1001 16:19:16.976919 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hq8t6" event={"ID":"551173b5-0f09-4b1e-831b-c6598ba904cb","Type":"ContainerDied","Data":"7dbe378442bb110761661bc29105a34b7e542487981a30e0ee6ff6c42928e6cd"} Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.323457 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.465440 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-config-data\") pod \"551173b5-0f09-4b1e-831b-c6598ba904cb\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.465751 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-combined-ca-bundle\") pod \"551173b5-0f09-4b1e-831b-c6598ba904cb\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.465946 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-scripts\") pod \"551173b5-0f09-4b1e-831b-c6598ba904cb\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.466108 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75v2t\" (UniqueName: \"kubernetes.io/projected/551173b5-0f09-4b1e-831b-c6598ba904cb-kube-api-access-75v2t\") pod \"551173b5-0f09-4b1e-831b-c6598ba904cb\" (UID: \"551173b5-0f09-4b1e-831b-c6598ba904cb\") " Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.477182 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-scripts" (OuterVolumeSpecName: "scripts") pod "551173b5-0f09-4b1e-831b-c6598ba904cb" (UID: "551173b5-0f09-4b1e-831b-c6598ba904cb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.477230 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/551173b5-0f09-4b1e-831b-c6598ba904cb-kube-api-access-75v2t" (OuterVolumeSpecName: "kube-api-access-75v2t") pod "551173b5-0f09-4b1e-831b-c6598ba904cb" (UID: "551173b5-0f09-4b1e-831b-c6598ba904cb"). InnerVolumeSpecName "kube-api-access-75v2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.496308 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-config-data" (OuterVolumeSpecName: "config-data") pod "551173b5-0f09-4b1e-831b-c6598ba904cb" (UID: "551173b5-0f09-4b1e-831b-c6598ba904cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.502717 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "551173b5-0f09-4b1e-831b-c6598ba904cb" (UID: "551173b5-0f09-4b1e-831b-c6598ba904cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.568572 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.568615 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75v2t\" (UniqueName: \"kubernetes.io/projected/551173b5-0f09-4b1e-831b-c6598ba904cb-kube-api-access-75v2t\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.568631 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.568644 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551173b5-0f09-4b1e-831b-c6598ba904cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.998464 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hq8t6" event={"ID":"551173b5-0f09-4b1e-831b-c6598ba904cb","Type":"ContainerDied","Data":"b7bb5e0b34a81f9494c7c74f8588c00e3036e68b71eee3d196d036d2e84d0c31"} Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.998516 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7bb5e0b34a81f9494c7c74f8588c00e3036e68b71eee3d196d036d2e84d0c31" Oct 01 16:19:18 crc kubenswrapper[4726]: I1001 16:19:18.998601 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hq8t6" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.105251 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 16:19:19 crc kubenswrapper[4726]: E1001 16:19:19.105619 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="551173b5-0f09-4b1e-831b-c6598ba904cb" containerName="nova-cell0-conductor-db-sync" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.105632 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="551173b5-0f09-4b1e-831b-c6598ba904cb" containerName="nova-cell0-conductor-db-sync" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.105821 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="551173b5-0f09-4b1e-831b-c6598ba904cb" containerName="nova-cell0-conductor-db-sync" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.106375 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.110022 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.110323 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-pdwp7" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.126712 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.181155 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db204841-b5e7-4ed3-bd88-e6fb7f9d627f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"db204841-b5e7-4ed3-bd88-e6fb7f9d627f\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.181734 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lb2l\" (UniqueName: \"kubernetes.io/projected/db204841-b5e7-4ed3-bd88-e6fb7f9d627f-kube-api-access-5lb2l\") pod \"nova-cell0-conductor-0\" (UID: \"db204841-b5e7-4ed3-bd88-e6fb7f9d627f\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.182030 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db204841-b5e7-4ed3-bd88-e6fb7f9d627f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"db204841-b5e7-4ed3-bd88-e6fb7f9d627f\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.284320 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lb2l\" (UniqueName: \"kubernetes.io/projected/db204841-b5e7-4ed3-bd88-e6fb7f9d627f-kube-api-access-5lb2l\") pod \"nova-cell0-conductor-0\" (UID: \"db204841-b5e7-4ed3-bd88-e6fb7f9d627f\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.284456 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db204841-b5e7-4ed3-bd88-e6fb7f9d627f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"db204841-b5e7-4ed3-bd88-e6fb7f9d627f\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.284492 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db204841-b5e7-4ed3-bd88-e6fb7f9d627f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"db204841-b5e7-4ed3-bd88-e6fb7f9d627f\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.294185 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db204841-b5e7-4ed3-bd88-e6fb7f9d627f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"db204841-b5e7-4ed3-bd88-e6fb7f9d627f\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.294278 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db204841-b5e7-4ed3-bd88-e6fb7f9d627f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"db204841-b5e7-4ed3-bd88-e6fb7f9d627f\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.301213 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lb2l\" (UniqueName: \"kubernetes.io/projected/db204841-b5e7-4ed3-bd88-e6fb7f9d627f-kube-api-access-5lb2l\") pod \"nova-cell0-conductor-0\" (UID: \"db204841-b5e7-4ed3-bd88-e6fb7f9d627f\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.438673 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 16:19:19 crc kubenswrapper[4726]: I1001 16:19:19.897622 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 16:19:20 crc kubenswrapper[4726]: I1001 16:19:20.020672 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"db204841-b5e7-4ed3-bd88-e6fb7f9d627f","Type":"ContainerStarted","Data":"7f9a8e449074c695953ca40f5c9f4afbbc0330466bb0c413628250c7315eb916"} Oct 01 16:19:21 crc kubenswrapper[4726]: I1001 16:19:21.037747 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"db204841-b5e7-4ed3-bd88-e6fb7f9d627f","Type":"ContainerStarted","Data":"cac901d1152a8295b3cbf847c24e8db633b8486c230513a954a4ff357e37dcc9"} Oct 01 16:19:21 crc kubenswrapper[4726]: I1001 16:19:21.038089 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 01 16:19:21 crc kubenswrapper[4726]: I1001 16:19:21.061358 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.061341895 podStartE2EDuration="2.061341895s" podCreationTimestamp="2025-10-01 16:19:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:19:21.055773088 +0000 UTC m=+1213.957325735" watchObservedRunningTime="2025-10-01 16:19:21.061341895 +0000 UTC m=+1213.962894472" Oct 01 16:19:23 crc kubenswrapper[4726]: I1001 16:19:23.175889 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 01 16:19:27 crc kubenswrapper[4726]: I1001 16:19:27.666644 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:19:27 crc kubenswrapper[4726]: I1001 16:19:27.667521 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="9639916d-aee7-485b-960f-c57fb48af4a0" containerName="kube-state-metrics" containerID="cri-o://b060292e5702871769507f89bf0f98f54b08a2b66040f7f1accfb5d82ed0d82a" gracePeriod=30 Oct 01 16:19:28 crc kubenswrapper[4726]: I1001 16:19:28.112402 4726 generic.go:334] "Generic (PLEG): container finished" podID="9639916d-aee7-485b-960f-c57fb48af4a0" containerID="b060292e5702871769507f89bf0f98f54b08a2b66040f7f1accfb5d82ed0d82a" exitCode=2 Oct 01 16:19:28 crc kubenswrapper[4726]: I1001 16:19:28.112447 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9639916d-aee7-485b-960f-c57fb48af4a0","Type":"ContainerDied","Data":"b060292e5702871769507f89bf0f98f54b08a2b66040f7f1accfb5d82ed0d82a"} Oct 01 16:19:28 crc kubenswrapper[4726]: I1001 16:19:28.345488 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 16:19:28 crc kubenswrapper[4726]: I1001 16:19:28.468218 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chp9b\" (UniqueName: \"kubernetes.io/projected/9639916d-aee7-485b-960f-c57fb48af4a0-kube-api-access-chp9b\") pod \"9639916d-aee7-485b-960f-c57fb48af4a0\" (UID: \"9639916d-aee7-485b-960f-c57fb48af4a0\") " Oct 01 16:19:28 crc kubenswrapper[4726]: I1001 16:19:28.489376 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9639916d-aee7-485b-960f-c57fb48af4a0-kube-api-access-chp9b" (OuterVolumeSpecName: "kube-api-access-chp9b") pod "9639916d-aee7-485b-960f-c57fb48af4a0" (UID: "9639916d-aee7-485b-960f-c57fb48af4a0"). InnerVolumeSpecName "kube-api-access-chp9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:19:28 crc kubenswrapper[4726]: I1001 16:19:28.570466 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chp9b\" (UniqueName: \"kubernetes.io/projected/9639916d-aee7-485b-960f-c57fb48af4a0-kube-api-access-chp9b\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.126457 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9639916d-aee7-485b-960f-c57fb48af4a0","Type":"ContainerDied","Data":"bab572d58892c3670a8e18157e982a146e27b07bbb4d25f66ae72e9834e67959"} Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.127881 4726 scope.go:117] "RemoveContainer" containerID="b060292e5702871769507f89bf0f98f54b08a2b66040f7f1accfb5d82ed0d82a" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.127022 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.168336 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.182985 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.211959 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:19:29 crc kubenswrapper[4726]: E1001 16:19:29.212384 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9639916d-aee7-485b-960f-c57fb48af4a0" containerName="kube-state-metrics" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.212404 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9639916d-aee7-485b-960f-c57fb48af4a0" containerName="kube-state-metrics" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.212574 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9639916d-aee7-485b-960f-c57fb48af4a0" containerName="kube-state-metrics" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.213218 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.215457 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.215763 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.236707 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.291081 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94r2w\" (UniqueName: \"kubernetes.io/projected/d17a343c-2ff2-4c6e-a224-c8d13f476243-kube-api-access-94r2w\") pod \"kube-state-metrics-0\" (UID: \"d17a343c-2ff2-4c6e-a224-c8d13f476243\") " pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.291286 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17a343c-2ff2-4c6e-a224-c8d13f476243-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d17a343c-2ff2-4c6e-a224-c8d13f476243\") " pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.291368 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d17a343c-2ff2-4c6e-a224-c8d13f476243-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d17a343c-2ff2-4c6e-a224-c8d13f476243\") " pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.291477 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d17a343c-2ff2-4c6e-a224-c8d13f476243-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d17a343c-2ff2-4c6e-a224-c8d13f476243\") " pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.393426 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17a343c-2ff2-4c6e-a224-c8d13f476243-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d17a343c-2ff2-4c6e-a224-c8d13f476243\") " pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.393477 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d17a343c-2ff2-4c6e-a224-c8d13f476243-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d17a343c-2ff2-4c6e-a224-c8d13f476243\") " pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.393511 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d17a343c-2ff2-4c6e-a224-c8d13f476243-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d17a343c-2ff2-4c6e-a224-c8d13f476243\") " pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.393596 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94r2w\" (UniqueName: \"kubernetes.io/projected/d17a343c-2ff2-4c6e-a224-c8d13f476243-kube-api-access-94r2w\") pod \"kube-state-metrics-0\" (UID: \"d17a343c-2ff2-4c6e-a224-c8d13f476243\") " pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.399807 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d17a343c-2ff2-4c6e-a224-c8d13f476243-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d17a343c-2ff2-4c6e-a224-c8d13f476243\") " pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.399973 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d17a343c-2ff2-4c6e-a224-c8d13f476243-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d17a343c-2ff2-4c6e-a224-c8d13f476243\") " pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.400445 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17a343c-2ff2-4c6e-a224-c8d13f476243-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d17a343c-2ff2-4c6e-a224-c8d13f476243\") " pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.425220 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94r2w\" (UniqueName: \"kubernetes.io/projected/d17a343c-2ff2-4c6e-a224-c8d13f476243-kube-api-access-94r2w\") pod \"kube-state-metrics-0\" (UID: \"d17a343c-2ff2-4c6e-a224-c8d13f476243\") " pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.472569 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.514779 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.515176 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="ceilometer-central-agent" containerID="cri-o://bb247c2c0d6727ac92bbefbe0691be261f8f09ea1f63430728df3189110da6a8" gracePeriod=30 Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.515265 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="sg-core" containerID="cri-o://b8e75a9ae62ebfe1b87a2db2d77ba2ee76f8485f8245d5668e777dee5acb3856" gracePeriod=30 Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.515329 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="ceilometer-notification-agent" containerID="cri-o://9e1d97f3da2ba2656715f91fa9077d540f2a066473452e65d17d17f4bb43094d" gracePeriod=30 Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.515280 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="proxy-httpd" containerID="cri-o://7c81d977f6de6cb974b9cf597dac3ff42f98c096b8992b13aaaf112c762e5487" gracePeriod=30 Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.538819 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 16:19:29 crc kubenswrapper[4726]: I1001 16:19:29.818707 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9639916d-aee7-485b-960f-c57fb48af4a0" path="/var/lib/kubelet/pods/9639916d-aee7-485b-960f-c57fb48af4a0/volumes" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.026760 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.106684 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-qggxs"] Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.108190 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.110751 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.111126 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.129989 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qggxs"] Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.170688 4726 generic.go:334] "Generic (PLEG): container finished" podID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerID="7c81d977f6de6cb974b9cf597dac3ff42f98c096b8992b13aaaf112c762e5487" exitCode=0 Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.170718 4726 generic.go:334] "Generic (PLEG): container finished" podID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerID="b8e75a9ae62ebfe1b87a2db2d77ba2ee76f8485f8245d5668e777dee5acb3856" exitCode=2 Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.170725 4726 generic.go:334] "Generic (PLEG): container finished" podID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerID="bb247c2c0d6727ac92bbefbe0691be261f8f09ea1f63430728df3189110da6a8" exitCode=0 Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.170769 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82fecdc7-09c8-4f55-aa63-5e953ada8783","Type":"ContainerDied","Data":"7c81d977f6de6cb974b9cf597dac3ff42f98c096b8992b13aaaf112c762e5487"} Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.170804 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82fecdc7-09c8-4f55-aa63-5e953ada8783","Type":"ContainerDied","Data":"b8e75a9ae62ebfe1b87a2db2d77ba2ee76f8485f8245d5668e777dee5acb3856"} Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.170814 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82fecdc7-09c8-4f55-aa63-5e953ada8783","Type":"ContainerDied","Data":"bb247c2c0d6727ac92bbefbe0691be261f8f09ea1f63430728df3189110da6a8"} Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.184449 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d17a343c-2ff2-4c6e-a224-c8d13f476243","Type":"ContainerStarted","Data":"2fce115e7447407447861b2d6bd4abf6020b862a3e4c7fd8c81ca48aad583283"} Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.211102 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwpb5\" (UniqueName: \"kubernetes.io/projected/9ee77a28-a1b7-4d80-9204-42a56d4dde89-kube-api-access-lwpb5\") pod \"nova-cell0-cell-mapping-qggxs\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.211225 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-config-data\") pod \"nova-cell0-cell-mapping-qggxs\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.211271 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-scripts\") pod \"nova-cell0-cell-mapping-qggxs\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.211341 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qggxs\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.249986 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.251094 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.254006 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.293323 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.314438 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-config-data\") pod \"nova-cell0-cell-mapping-qggxs\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.314525 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.314601 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-scripts\") pod \"nova-cell0-cell-mapping-qggxs\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.314744 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qggxs\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.314781 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66g59\" (UniqueName: \"kubernetes.io/projected/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-kube-api-access-66g59\") pod \"nova-scheduler-0\" (UID: \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.314848 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwpb5\" (UniqueName: \"kubernetes.io/projected/9ee77a28-a1b7-4d80-9204-42a56d4dde89-kube-api-access-lwpb5\") pod \"nova-cell0-cell-mapping-qggxs\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.314887 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-config-data\") pod \"nova-scheduler-0\" (UID: \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.333463 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-config-data\") pod \"nova-cell0-cell-mapping-qggxs\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.334768 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qggxs\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.337556 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-scripts\") pod \"nova-cell0-cell-mapping-qggxs\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.350849 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.352821 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.356392 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.378857 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwpb5\" (UniqueName: \"kubernetes.io/projected/9ee77a28-a1b7-4d80-9204-42a56d4dde89-kube-api-access-lwpb5\") pod \"nova-cell0-cell-mapping-qggxs\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.381170 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.419587 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-config-data\") pod \"nova-scheduler-0\" (UID: \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.419663 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3511a2-5943-4017-9c11-c42ac4bfa379-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " pod="openstack/nova-metadata-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.419724 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a3511a2-5943-4017-9c11-c42ac4bfa379-logs\") pod \"nova-metadata-0\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " pod="openstack/nova-metadata-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.419761 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3511a2-5943-4017-9c11-c42ac4bfa379-config-data\") pod \"nova-metadata-0\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " pod="openstack/nova-metadata-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.419825 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.419925 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66g59\" (UniqueName: \"kubernetes.io/projected/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-kube-api-access-66g59\") pod \"nova-scheduler-0\" (UID: \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.419975 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h56mt\" (UniqueName: \"kubernetes.io/projected/2a3511a2-5943-4017-9c11-c42ac4bfa379-kube-api-access-h56mt\") pod \"nova-metadata-0\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " pod="openstack/nova-metadata-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.436882 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.437063 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-config-data\") pod \"nova-scheduler-0\" (UID: \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.439636 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.444851 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.458169 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66g59\" (UniqueName: \"kubernetes.io/projected/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-kube-api-access-66g59\") pod \"nova-scheduler-0\" (UID: \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.470458 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.474448 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.481161 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.521336 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a3511a2-5943-4017-9c11-c42ac4bfa379-logs\") pod \"nova-metadata-0\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " pod="openstack/nova-metadata-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.521378 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3511a2-5943-4017-9c11-c42ac4bfa379-config-data\") pod \"nova-metadata-0\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " pod="openstack/nova-metadata-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.521494 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h56mt\" (UniqueName: \"kubernetes.io/projected/2a3511a2-5943-4017-9c11-c42ac4bfa379-kube-api-access-h56mt\") pod \"nova-metadata-0\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " pod="openstack/nova-metadata-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.521530 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3511a2-5943-4017-9c11-c42ac4bfa379-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " pod="openstack/nova-metadata-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.527740 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-md5d8"] Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.528099 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a3511a2-5943-4017-9c11-c42ac4bfa379-logs\") pod \"nova-metadata-0\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " pod="openstack/nova-metadata-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.528572 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3511a2-5943-4017-9c11-c42ac4bfa379-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " pod="openstack/nova-metadata-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.529117 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3511a2-5943-4017-9c11-c42ac4bfa379-config-data\") pod \"nova-metadata-0\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " pod="openstack/nova-metadata-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.583474 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.611853 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.623804 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h56mt\" (UniqueName: \"kubernetes.io/projected/2a3511a2-5943-4017-9c11-c42ac4bfa379-kube-api-access-h56mt\") pod \"nova-metadata-0\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " pod="openstack/nova-metadata-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.653724 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55ae01da-8b6c-464b-b77b-5f0358fd8b18-logs\") pod \"nova-api-0\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " pod="openstack/nova-api-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.653856 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ae01da-8b6c-464b-b77b-5f0358fd8b18-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " pod="openstack/nova-api-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.653926 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ae01da-8b6c-464b-b77b-5f0358fd8b18-config-data\") pod \"nova-api-0\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " pod="openstack/nova-api-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.653968 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjx2r\" (UniqueName: \"kubernetes.io/projected/55ae01da-8b6c-464b-b77b-5f0358fd8b18-kube-api-access-rjx2r\") pod \"nova-api-0\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " pod="openstack/nova-api-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.669639 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-md5d8"] Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.683420 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.702235 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.703509 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.705295 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.746896 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.765457 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55ae01da-8b6c-464b-b77b-5f0358fd8b18-logs\") pod \"nova-api-0\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " pod="openstack/nova-api-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.765510 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.765544 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ae01da-8b6c-464b-b77b-5f0358fd8b18-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " pod="openstack/nova-api-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.765573 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ae01da-8b6c-464b-b77b-5f0358fd8b18-config-data\") pod \"nova-api-0\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " pod="openstack/nova-api-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.765595 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-dns-svc\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.765611 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjx2r\" (UniqueName: \"kubernetes.io/projected/55ae01da-8b6c-464b-b77b-5f0358fd8b18-kube-api-access-rjx2r\") pod \"nova-api-0\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " pod="openstack/nova-api-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.765635 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmv5x\" (UniqueName: \"kubernetes.io/projected/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-kube-api-access-nmv5x\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.765690 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-config\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.765718 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.765739 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.766175 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55ae01da-8b6c-464b-b77b-5f0358fd8b18-logs\") pod \"nova-api-0\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " pod="openstack/nova-api-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.778429 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ae01da-8b6c-464b-b77b-5f0358fd8b18-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " pod="openstack/nova-api-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.789765 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjx2r\" (UniqueName: \"kubernetes.io/projected/55ae01da-8b6c-464b-b77b-5f0358fd8b18-kube-api-access-rjx2r\") pod \"nova-api-0\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " pod="openstack/nova-api-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.802578 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ae01da-8b6c-464b-b77b-5f0358fd8b18-config-data\") pod \"nova-api-0\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " pod="openstack/nova-api-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.868290 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-config\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.868537 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.868646 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.868731 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95gwn\" (UniqueName: \"kubernetes.io/projected/dd6f8ebd-b1b9-4142-8012-4310c45a041f-kube-api-access-95gwn\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.868808 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6f8ebd-b1b9-4142-8012-4310c45a041f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.868911 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.869027 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-dns-svc\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.869142 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmv5x\" (UniqueName: \"kubernetes.io/projected/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-kube-api-access-nmv5x\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.869759 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.869842 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.869936 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-config\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.870107 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6f8ebd-b1b9-4142-8012-4310c45a041f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.870859 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-dns-svc\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.871320 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.894519 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmv5x\" (UniqueName: \"kubernetes.io/projected/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-kube-api-access-nmv5x\") pod \"dnsmasq-dns-bccf8f775-md5d8\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.971421 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.973759 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95gwn\" (UniqueName: \"kubernetes.io/projected/dd6f8ebd-b1b9-4142-8012-4310c45a041f-kube-api-access-95gwn\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.973811 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6f8ebd-b1b9-4142-8012-4310c45a041f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.973988 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6f8ebd-b1b9-4142-8012-4310c45a041f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.979987 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6f8ebd-b1b9-4142-8012-4310c45a041f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.986039 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6f8ebd-b1b9-4142-8012-4310c45a041f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.992540 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95gwn\" (UniqueName: \"kubernetes.io/projected/dd6f8ebd-b1b9-4142-8012-4310c45a041f-kube-api-access-95gwn\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:19:30 crc kubenswrapper[4726]: I1001 16:19:30.996386 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.048912 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.132016 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qggxs"] Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.198359 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qggxs" event={"ID":"9ee77a28-a1b7-4d80-9204-42a56d4dde89","Type":"ContainerStarted","Data":"f0cf93ab7aee5a7f7343f73503492ce0edfe145e526e632a1c33c343d0a5bd87"} Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.200353 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.292206 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.574206 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-79vp9"] Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.575743 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.585891 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-79vp9"] Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.586509 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.586706 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.658906 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-md5d8"] Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.689096 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-config-data\") pod \"nova-cell1-conductor-db-sync-79vp9\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.689139 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9bfs\" (UniqueName: \"kubernetes.io/projected/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-kube-api-access-s9bfs\") pod \"nova-cell1-conductor-db-sync-79vp9\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.689256 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-scripts\") pod \"nova-cell1-conductor-db-sync-79vp9\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.689274 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-79vp9\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.729621 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.769838 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.790884 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-scripts\") pod \"nova-cell1-conductor-db-sync-79vp9\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.790925 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-79vp9\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.790974 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-config-data\") pod \"nova-cell1-conductor-db-sync-79vp9\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.790994 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9bfs\" (UniqueName: \"kubernetes.io/projected/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-kube-api-access-s9bfs\") pod \"nova-cell1-conductor-db-sync-79vp9\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.795302 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-config-data\") pod \"nova-cell1-conductor-db-sync-79vp9\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.795297 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-79vp9\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.796988 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-scripts\") pod \"nova-cell1-conductor-db-sync-79vp9\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.816569 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9bfs\" (UniqueName: \"kubernetes.io/projected/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-kube-api-access-s9bfs\") pod \"nova-cell1-conductor-db-sync-79vp9\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:31 crc kubenswrapper[4726]: I1001 16:19:31.909317 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.222222 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6","Type":"ContainerStarted","Data":"b50cb0626bb1f0aac79c3b8bc58e97c88ed1975a21790d2c9c3f53b05e61e2d9"} Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.224357 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d17a343c-2ff2-4c6e-a224-c8d13f476243","Type":"ContainerStarted","Data":"ca81d59d7e50c797d0e49a9554eea2cda694c3ccb1beeee57d5078062d8a4747"} Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.226398 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.243982 4726 generic.go:334] "Generic (PLEG): container finished" podID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerID="9e1d97f3da2ba2656715f91fa9077d540f2a066473452e65d17d17f4bb43094d" exitCode=0 Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.244240 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82fecdc7-09c8-4f55-aa63-5e953ada8783","Type":"ContainerDied","Data":"9e1d97f3da2ba2656715f91fa9077d540f2a066473452e65d17d17f4bb43094d"} Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.248327 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.3753434860000002 podStartE2EDuration="3.248308538s" podCreationTimestamp="2025-10-01 16:19:29 +0000 UTC" firstStartedPulling="2025-10-01 16:19:30.028829847 +0000 UTC m=+1222.930382464" lastFinishedPulling="2025-10-01 16:19:30.901794939 +0000 UTC m=+1223.803347516" observedRunningTime="2025-10-01 16:19:32.244570113 +0000 UTC m=+1225.146122700" watchObservedRunningTime="2025-10-01 16:19:32.248308538 +0000 UTC m=+1225.149861115" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.254147 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dd6f8ebd-b1b9-4142-8012-4310c45a041f","Type":"ContainerStarted","Data":"f1f258a750a7d790beb723b9fb3e124727bf7f91340a474e6c8de7b63034bfba"} Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.257277 4726 generic.go:334] "Generic (PLEG): container finished" podID="60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" containerID="38c5a02a0e7abe05039101d8539cf994fff94889411917ed1bc64191d7442572" exitCode=0 Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.257375 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-md5d8" event={"ID":"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6","Type":"ContainerDied","Data":"38c5a02a0e7abe05039101d8539cf994fff94889411917ed1bc64191d7442572"} Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.257458 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-md5d8" event={"ID":"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6","Type":"ContainerStarted","Data":"086eec88302b30529260a6b184e4c77826e6178cb3069fc4a7ddf8158a4b7ea8"} Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.261137 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2a3511a2-5943-4017-9c11-c42ac4bfa379","Type":"ContainerStarted","Data":"a6d4838c6d0c52ac1d48664a996182de013bdff2dc39a81497f621bf6968bd23"} Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.276652 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qggxs" event={"ID":"9ee77a28-a1b7-4d80-9204-42a56d4dde89","Type":"ContainerStarted","Data":"04aa3a2c1c842a1fb997a6f412f4d5e2fa43fbf04dc54e8452e295344b5cf5ca"} Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.279555 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55ae01da-8b6c-464b-b77b-5f0358fd8b18","Type":"ContainerStarted","Data":"e52d91953a8f2dcf15e046bac96c3a65f74de10adbfc64e1e6583954961a08ee"} Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.324339 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-qggxs" podStartSLOduration=2.324317521 podStartE2EDuration="2.324317521s" podCreationTimestamp="2025-10-01 16:19:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:19:32.306660133 +0000 UTC m=+1225.208212730" watchObservedRunningTime="2025-10-01 16:19:32.324317521 +0000 UTC m=+1225.225870098" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.471357 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-79vp9"] Oct 01 16:19:32 crc kubenswrapper[4726]: W1001 16:19:32.491707 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd747a1f6_2d98_4b9e_9a0d_833d0afede9d.slice/crio-9c7c00e6c80b563fd49bc8e33e4022b93e4376d6a9709d2b1ed0621acc9e8c58 WatchSource:0}: Error finding container 9c7c00e6c80b563fd49bc8e33e4022b93e4376d6a9709d2b1ed0621acc9e8c58: Status 404 returned error can't find the container with id 9c7c00e6c80b563fd49bc8e33e4022b93e4376d6a9709d2b1ed0621acc9e8c58 Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.716490 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.822737 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-sg-core-conf-yaml\") pod \"82fecdc7-09c8-4f55-aa63-5e953ada8783\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.822804 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-scripts\") pod \"82fecdc7-09c8-4f55-aa63-5e953ada8783\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.822840 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-combined-ca-bundle\") pod \"82fecdc7-09c8-4f55-aa63-5e953ada8783\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.822886 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-config-data\") pod \"82fecdc7-09c8-4f55-aa63-5e953ada8783\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.822988 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrtj8\" (UniqueName: \"kubernetes.io/projected/82fecdc7-09c8-4f55-aa63-5e953ada8783-kube-api-access-rrtj8\") pod \"82fecdc7-09c8-4f55-aa63-5e953ada8783\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.823036 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82fecdc7-09c8-4f55-aa63-5e953ada8783-log-httpd\") pod \"82fecdc7-09c8-4f55-aa63-5e953ada8783\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.823134 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82fecdc7-09c8-4f55-aa63-5e953ada8783-run-httpd\") pod \"82fecdc7-09c8-4f55-aa63-5e953ada8783\" (UID: \"82fecdc7-09c8-4f55-aa63-5e953ada8783\") " Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.823864 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82fecdc7-09c8-4f55-aa63-5e953ada8783-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "82fecdc7-09c8-4f55-aa63-5e953ada8783" (UID: "82fecdc7-09c8-4f55-aa63-5e953ada8783"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.824812 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82fecdc7-09c8-4f55-aa63-5e953ada8783-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "82fecdc7-09c8-4f55-aa63-5e953ada8783" (UID: "82fecdc7-09c8-4f55-aa63-5e953ada8783"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.830166 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82fecdc7-09c8-4f55-aa63-5e953ada8783-kube-api-access-rrtj8" (OuterVolumeSpecName: "kube-api-access-rrtj8") pod "82fecdc7-09c8-4f55-aa63-5e953ada8783" (UID: "82fecdc7-09c8-4f55-aa63-5e953ada8783"). InnerVolumeSpecName "kube-api-access-rrtj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.834288 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-scripts" (OuterVolumeSpecName: "scripts") pod "82fecdc7-09c8-4f55-aa63-5e953ada8783" (UID: "82fecdc7-09c8-4f55-aa63-5e953ada8783"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.901961 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "82fecdc7-09c8-4f55-aa63-5e953ada8783" (UID: "82fecdc7-09c8-4f55-aa63-5e953ada8783"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.928455 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrtj8\" (UniqueName: \"kubernetes.io/projected/82fecdc7-09c8-4f55-aa63-5e953ada8783-kube-api-access-rrtj8\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.928691 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82fecdc7-09c8-4f55-aa63-5e953ada8783-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.928708 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82fecdc7-09c8-4f55-aa63-5e953ada8783-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.928719 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.928731 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:32 crc kubenswrapper[4726]: I1001 16:19:32.981503 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-config-data" (OuterVolumeSpecName: "config-data") pod "82fecdc7-09c8-4f55-aa63-5e953ada8783" (UID: "82fecdc7-09c8-4f55-aa63-5e953ada8783"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.009171 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82fecdc7-09c8-4f55-aa63-5e953ada8783" (UID: "82fecdc7-09c8-4f55-aa63-5e953ada8783"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.044849 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.044889 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82fecdc7-09c8-4f55-aa63-5e953ada8783-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.311102 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82fecdc7-09c8-4f55-aa63-5e953ada8783","Type":"ContainerDied","Data":"3906fa24eb4b6beb55bed80efb86ecba4be6d4c19214d1b19a98e9eee013cb63"} Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.311164 4726 scope.go:117] "RemoveContainer" containerID="7c81d977f6de6cb974b9cf597dac3ff42f98c096b8992b13aaaf112c762e5487" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.311229 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.317925 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-md5d8" event={"ID":"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6","Type":"ContainerStarted","Data":"1b0d9345e2119c42c63000a1a380db0f692b70f9faee7873767ae62a1654c858"} Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.319427 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.323297 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-79vp9" event={"ID":"d747a1f6-2d98-4b9e-9a0d-833d0afede9d","Type":"ContainerStarted","Data":"ed81401bbb74ca249f20375c440e14f84f09d3a54734aef9dc6aed10e691944d"} Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.323329 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-79vp9" event={"ID":"d747a1f6-2d98-4b9e-9a0d-833d0afede9d","Type":"ContainerStarted","Data":"9c7c00e6c80b563fd49bc8e33e4022b93e4376d6a9709d2b1ed0621acc9e8c58"} Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.358629 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-md5d8" podStartSLOduration=3.3586089599999998 podStartE2EDuration="3.35860896s" podCreationTimestamp="2025-10-01 16:19:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:19:33.338948376 +0000 UTC m=+1226.240500963" watchObservedRunningTime="2025-10-01 16:19:33.35860896 +0000 UTC m=+1226.260161537" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.378887 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.394126 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.398012 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-79vp9" podStartSLOduration=2.39798651 podStartE2EDuration="2.39798651s" podCreationTimestamp="2025-10-01 16:19:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:19:33.376967368 +0000 UTC m=+1226.278519945" watchObservedRunningTime="2025-10-01 16:19:33.39798651 +0000 UTC m=+1226.299539097" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.410131 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:19:33 crc kubenswrapper[4726]: E1001 16:19:33.410849 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="proxy-httpd" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.410865 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="proxy-httpd" Oct 01 16:19:33 crc kubenswrapper[4726]: E1001 16:19:33.410890 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="ceilometer-notification-agent" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.410898 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="ceilometer-notification-agent" Oct 01 16:19:33 crc kubenswrapper[4726]: E1001 16:19:33.410913 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="ceilometer-central-agent" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.410921 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="ceilometer-central-agent" Oct 01 16:19:33 crc kubenswrapper[4726]: E1001 16:19:33.410933 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="sg-core" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.410941 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="sg-core" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.411137 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="ceilometer-central-agent" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.411151 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="proxy-httpd" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.411164 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="ceilometer-notification-agent" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.411184 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" containerName="sg-core" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.413065 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.416571 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.416761 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.416957 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.419588 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.563623 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fdf946f7-e990-4849-a998-3d151a619bf2-log-httpd\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.563706 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.563745 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sfp9\" (UniqueName: \"kubernetes.io/projected/fdf946f7-e990-4849-a998-3d151a619bf2-kube-api-access-7sfp9\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.563839 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-config-data\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.564979 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.565021 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-scripts\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.565400 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fdf946f7-e990-4849-a998-3d151a619bf2-run-httpd\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.565485 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.666770 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fdf946f7-e990-4849-a998-3d151a619bf2-run-httpd\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.666826 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.666901 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fdf946f7-e990-4849-a998-3d151a619bf2-log-httpd\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.666950 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.666990 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sfp9\" (UniqueName: \"kubernetes.io/projected/fdf946f7-e990-4849-a998-3d151a619bf2-kube-api-access-7sfp9\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.667038 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-config-data\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.667091 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.667123 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-scripts\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.670665 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fdf946f7-e990-4849-a998-3d151a619bf2-run-httpd\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.670908 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fdf946f7-e990-4849-a998-3d151a619bf2-log-httpd\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.675617 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-scripts\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.677479 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.677894 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.678283 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-config-data\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.680622 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.687687 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sfp9\" (UniqueName: \"kubernetes.io/projected/fdf946f7-e990-4849-a998-3d151a619bf2-kube-api-access-7sfp9\") pod \"ceilometer-0\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.773325 4726 scope.go:117] "RemoveContainer" containerID="b8e75a9ae62ebfe1b87a2db2d77ba2ee76f8485f8245d5668e777dee5acb3856" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.795819 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:19:33 crc kubenswrapper[4726]: I1001 16:19:33.818188 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82fecdc7-09c8-4f55-aa63-5e953ada8783" path="/var/lib/kubelet/pods/82fecdc7-09c8-4f55-aa63-5e953ada8783/volumes" Oct 01 16:19:34 crc kubenswrapper[4726]: I1001 16:19:34.612737 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:19:34 crc kubenswrapper[4726]: I1001 16:19:34.625696 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:36 crc kubenswrapper[4726]: I1001 16:19:36.551559 4726 scope.go:117] "RemoveContainer" containerID="9e1d97f3da2ba2656715f91fa9077d540f2a066473452e65d17d17f4bb43094d" Oct 01 16:19:37 crc kubenswrapper[4726]: I1001 16:19:37.134537 4726 scope.go:117] "RemoveContainer" containerID="bb247c2c0d6727ac92bbefbe0691be261f8f09ea1f63430728df3189110da6a8" Oct 01 16:19:37 crc kubenswrapper[4726]: W1001 16:19:37.620022 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdf946f7_e990_4849_a998_3d151a619bf2.slice/crio-fb4ec01d0824f3dd1bd3fb39cc5f104318934745df1c8604a1f0cc976ecf2485 WatchSource:0}: Error finding container fb4ec01d0824f3dd1bd3fb39cc5f104318934745df1c8604a1f0cc976ecf2485: Status 404 returned error can't find the container with id fb4ec01d0824f3dd1bd3fb39cc5f104318934745df1c8604a1f0cc976ecf2485 Oct 01 16:19:37 crc kubenswrapper[4726]: I1001 16:19:37.624777 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.393208 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dd6f8ebd-b1b9-4142-8012-4310c45a041f","Type":"ContainerStarted","Data":"7c094296cbbddf544b6fd7851462d7a082546e3a64eec89121444d8f1dcbcccf"} Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.393284 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="dd6f8ebd-b1b9-4142-8012-4310c45a041f" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://7c094296cbbddf544b6fd7851462d7a082546e3a64eec89121444d8f1dcbcccf" gracePeriod=30 Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.400284 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2a3511a2-5943-4017-9c11-c42ac4bfa379","Type":"ContainerStarted","Data":"26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2"} Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.400327 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2a3511a2-5943-4017-9c11-c42ac4bfa379","Type":"ContainerStarted","Data":"5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d"} Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.400442 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2a3511a2-5943-4017-9c11-c42ac4bfa379" containerName="nova-metadata-log" containerID="cri-o://5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d" gracePeriod=30 Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.400745 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2a3511a2-5943-4017-9c11-c42ac4bfa379" containerName="nova-metadata-metadata" containerID="cri-o://26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2" gracePeriod=30 Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.408498 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.063817841 podStartE2EDuration="8.408479941s" podCreationTimestamp="2025-10-01 16:19:30 +0000 UTC" firstStartedPulling="2025-10-01 16:19:31.833026634 +0000 UTC m=+1224.734579211" lastFinishedPulling="2025-10-01 16:19:37.177688724 +0000 UTC m=+1230.079241311" observedRunningTime="2025-10-01 16:19:38.406475504 +0000 UTC m=+1231.308028081" watchObservedRunningTime="2025-10-01 16:19:38.408479941 +0000 UTC m=+1231.310032518" Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.414488 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55ae01da-8b6c-464b-b77b-5f0358fd8b18","Type":"ContainerStarted","Data":"fec82a33896f952be8d9a59bf54762dc24e72dca6807e461b2eca2d18efb2581"} Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.414537 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55ae01da-8b6c-464b-b77b-5f0358fd8b18","Type":"ContainerStarted","Data":"e96fa9dba30321a23befb6390facdced1fd809bfae6624f8d38080a92a72d9d6"} Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.416970 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6","Type":"ContainerStarted","Data":"dcb394db92f47b55646f5397129b6f8bcf575fe56c9d6b29dc7050e0cb1a56fc"} Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.419176 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fdf946f7-e990-4849-a998-3d151a619bf2","Type":"ContainerStarted","Data":"fb4ec01d0824f3dd1bd3fb39cc5f104318934745df1c8604a1f0cc976ecf2485"} Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.429182 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.5615400040000003 podStartE2EDuration="8.429161413s" podCreationTimestamp="2025-10-01 16:19:30 +0000 UTC" firstStartedPulling="2025-10-01 16:19:31.31662518 +0000 UTC m=+1224.218177757" lastFinishedPulling="2025-10-01 16:19:37.184246589 +0000 UTC m=+1230.085799166" observedRunningTime="2025-10-01 16:19:38.423443002 +0000 UTC m=+1231.324995579" watchObservedRunningTime="2025-10-01 16:19:38.429161413 +0000 UTC m=+1231.330714000" Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.451720 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.986612545 podStartE2EDuration="8.451701259s" podCreationTimestamp="2025-10-01 16:19:30 +0000 UTC" firstStartedPulling="2025-10-01 16:19:31.735444724 +0000 UTC m=+1224.636997291" lastFinishedPulling="2025-10-01 16:19:37.200533428 +0000 UTC m=+1230.102086005" observedRunningTime="2025-10-01 16:19:38.443732074 +0000 UTC m=+1231.345284661" watchObservedRunningTime="2025-10-01 16:19:38.451701259 +0000 UTC m=+1231.353253836" Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.467006 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.508125429 podStartE2EDuration="8.466984499s" podCreationTimestamp="2025-10-01 16:19:30 +0000 UTC" firstStartedPulling="2025-10-01 16:19:31.219008109 +0000 UTC m=+1224.120560686" lastFinishedPulling="2025-10-01 16:19:37.177867179 +0000 UTC m=+1230.079419756" observedRunningTime="2025-10-01 16:19:38.457227954 +0000 UTC m=+1231.358780531" watchObservedRunningTime="2025-10-01 16:19:38.466984499 +0000 UTC m=+1231.368537076" Oct 01 16:19:38 crc kubenswrapper[4726]: I1001 16:19:38.917628 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.001126 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h56mt\" (UniqueName: \"kubernetes.io/projected/2a3511a2-5943-4017-9c11-c42ac4bfa379-kube-api-access-h56mt\") pod \"2a3511a2-5943-4017-9c11-c42ac4bfa379\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.001284 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3511a2-5943-4017-9c11-c42ac4bfa379-config-data\") pod \"2a3511a2-5943-4017-9c11-c42ac4bfa379\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.001369 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a3511a2-5943-4017-9c11-c42ac4bfa379-logs\") pod \"2a3511a2-5943-4017-9c11-c42ac4bfa379\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.001463 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3511a2-5943-4017-9c11-c42ac4bfa379-combined-ca-bundle\") pod \"2a3511a2-5943-4017-9c11-c42ac4bfa379\" (UID: \"2a3511a2-5943-4017-9c11-c42ac4bfa379\") " Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.002272 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a3511a2-5943-4017-9c11-c42ac4bfa379-logs" (OuterVolumeSpecName: "logs") pod "2a3511a2-5943-4017-9c11-c42ac4bfa379" (UID: "2a3511a2-5943-4017-9c11-c42ac4bfa379"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.002407 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a3511a2-5943-4017-9c11-c42ac4bfa379-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.012103 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a3511a2-5943-4017-9c11-c42ac4bfa379-kube-api-access-h56mt" (OuterVolumeSpecName: "kube-api-access-h56mt") pod "2a3511a2-5943-4017-9c11-c42ac4bfa379" (UID: "2a3511a2-5943-4017-9c11-c42ac4bfa379"). InnerVolumeSpecName "kube-api-access-h56mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.031822 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a3511a2-5943-4017-9c11-c42ac4bfa379-config-data" (OuterVolumeSpecName: "config-data") pod "2a3511a2-5943-4017-9c11-c42ac4bfa379" (UID: "2a3511a2-5943-4017-9c11-c42ac4bfa379"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.050238 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a3511a2-5943-4017-9c11-c42ac4bfa379-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a3511a2-5943-4017-9c11-c42ac4bfa379" (UID: "2a3511a2-5943-4017-9c11-c42ac4bfa379"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.104517 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h56mt\" (UniqueName: \"kubernetes.io/projected/2a3511a2-5943-4017-9c11-c42ac4bfa379-kube-api-access-h56mt\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.104557 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a3511a2-5943-4017-9c11-c42ac4bfa379-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.104569 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3511a2-5943-4017-9c11-c42ac4bfa379-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.429361 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fdf946f7-e990-4849-a998-3d151a619bf2","Type":"ContainerStarted","Data":"b9abcb235c3f3a08d5e9ad4619b5cba68560ecc4bb9c612636baba38cee061be"} Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.431045 4726 generic.go:334] "Generic (PLEG): container finished" podID="2a3511a2-5943-4017-9c11-c42ac4bfa379" containerID="26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2" exitCode=0 Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.431104 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2a3511a2-5943-4017-9c11-c42ac4bfa379","Type":"ContainerDied","Data":"26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2"} Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.431124 4726 generic.go:334] "Generic (PLEG): container finished" podID="2a3511a2-5943-4017-9c11-c42ac4bfa379" containerID="5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d" exitCode=143 Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.431151 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2a3511a2-5943-4017-9c11-c42ac4bfa379","Type":"ContainerDied","Data":"5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d"} Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.431165 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.431176 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2a3511a2-5943-4017-9c11-c42ac4bfa379","Type":"ContainerDied","Data":"a6d4838c6d0c52ac1d48664a996182de013bdff2dc39a81497f621bf6968bd23"} Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.431196 4726 scope.go:117] "RemoveContainer" containerID="26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.432381 4726 generic.go:334] "Generic (PLEG): container finished" podID="9ee77a28-a1b7-4d80-9204-42a56d4dde89" containerID="04aa3a2c1c842a1fb997a6f412f4d5e2fa43fbf04dc54e8452e295344b5cf5ca" exitCode=0 Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.432459 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qggxs" event={"ID":"9ee77a28-a1b7-4d80-9204-42a56d4dde89","Type":"ContainerDied","Data":"04aa3a2c1c842a1fb997a6f412f4d5e2fa43fbf04dc54e8452e295344b5cf5ca"} Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.489891 4726 scope.go:117] "RemoveContainer" containerID="5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.515082 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.528585 4726 scope.go:117] "RemoveContainer" containerID="26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2" Oct 01 16:19:39 crc kubenswrapper[4726]: E1001 16:19:39.529169 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2\": container with ID starting with 26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2 not found: ID does not exist" containerID="26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.529224 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2"} err="failed to get container status \"26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2\": rpc error: code = NotFound desc = could not find container \"26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2\": container with ID starting with 26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2 not found: ID does not exist" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.529256 4726 scope.go:117] "RemoveContainer" containerID="5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d" Oct 01 16:19:39 crc kubenswrapper[4726]: E1001 16:19:39.529673 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d\": container with ID starting with 5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d not found: ID does not exist" containerID="5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.529715 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d"} err="failed to get container status \"5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d\": rpc error: code = NotFound desc = could not find container \"5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d\": container with ID starting with 5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d not found: ID does not exist" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.529742 4726 scope.go:117] "RemoveContainer" containerID="26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.534259 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2"} err="failed to get container status \"26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2\": rpc error: code = NotFound desc = could not find container \"26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2\": container with ID starting with 26530cb411713367824ddb98cde2a7ad18dd7b2466d712cdc0f1e43bd94d0ef2 not found: ID does not exist" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.534376 4726 scope.go:117] "RemoveContainer" containerID="5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.539885 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.540577 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d"} err="failed to get container status \"5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d\": rpc error: code = NotFound desc = could not find container \"5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d\": container with ID starting with 5cdc8116f904a6617c862c9653b73a33ff9e6a3d3b80432034a293e3a2618a8d not found: ID does not exist" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.555824 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:39 crc kubenswrapper[4726]: E1001 16:19:39.556261 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a3511a2-5943-4017-9c11-c42ac4bfa379" containerName="nova-metadata-log" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.556282 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a3511a2-5943-4017-9c11-c42ac4bfa379" containerName="nova-metadata-log" Oct 01 16:19:39 crc kubenswrapper[4726]: E1001 16:19:39.556318 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a3511a2-5943-4017-9c11-c42ac4bfa379" containerName="nova-metadata-metadata" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.556335 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a3511a2-5943-4017-9c11-c42ac4bfa379" containerName="nova-metadata-metadata" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.556558 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a3511a2-5943-4017-9c11-c42ac4bfa379" containerName="nova-metadata-metadata" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.556601 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a3511a2-5943-4017-9c11-c42ac4bfa379" containerName="nova-metadata-log" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.557822 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.561934 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.562253 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.564428 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.574577 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.612972 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9906cce7-2c94-4290-971e-d8a556a3a745-logs\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.613054 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-config-data\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.613169 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.613194 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9zk4\" (UniqueName: \"kubernetes.io/projected/9906cce7-2c94-4290-971e-d8a556a3a745-kube-api-access-s9zk4\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.613272 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.714847 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.715790 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9906cce7-2c94-4290-971e-d8a556a3a745-logs\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.715837 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-config-data\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.715923 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.715952 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9zk4\" (UniqueName: \"kubernetes.io/projected/9906cce7-2c94-4290-971e-d8a556a3a745-kube-api-access-s9zk4\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.716332 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9906cce7-2c94-4290-971e-d8a556a3a745-logs\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.719886 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.719938 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-config-data\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.722659 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.732031 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9zk4\" (UniqueName: \"kubernetes.io/projected/9906cce7-2c94-4290-971e-d8a556a3a745-kube-api-access-s9zk4\") pod \"nova-metadata-0\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " pod="openstack/nova-metadata-0" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.828495 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a3511a2-5943-4017-9c11-c42ac4bfa379" path="/var/lib/kubelet/pods/2a3511a2-5943-4017-9c11-c42ac4bfa379/volumes" Oct 01 16:19:39 crc kubenswrapper[4726]: I1001 16:19:39.880997 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:19:40 crc kubenswrapper[4726]: W1001 16:19:40.414227 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9906cce7_2c94_4290_971e_d8a556a3a745.slice/crio-9eba22707533e9fc27488ac5b6b69640ab3dcd0d9d3ba29e039bf417407bc181 WatchSource:0}: Error finding container 9eba22707533e9fc27488ac5b6b69640ab3dcd0d9d3ba29e039bf417407bc181: Status 404 returned error can't find the container with id 9eba22707533e9fc27488ac5b6b69640ab3dcd0d9d3ba29e039bf417407bc181 Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.418738 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.447687 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fdf946f7-e990-4849-a998-3d151a619bf2","Type":"ContainerStarted","Data":"602e5266a49092b8ddb435ec2c1bc3d6c770ccbd5be1058f3371c6c20f34fd93"} Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.447749 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fdf946f7-e990-4849-a998-3d151a619bf2","Type":"ContainerStarted","Data":"a0f250ba81efd279b348c1a80272ccb11c648ac1542741ec2633d1bb5eed421b"} Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.453347 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9906cce7-2c94-4290-971e-d8a556a3a745","Type":"ContainerStarted","Data":"9eba22707533e9fc27488ac5b6b69640ab3dcd0d9d3ba29e039bf417407bc181"} Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.613685 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.614195 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.664761 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.761384 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.840590 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-combined-ca-bundle\") pod \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.840815 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-scripts\") pod \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.840863 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwpb5\" (UniqueName: \"kubernetes.io/projected/9ee77a28-a1b7-4d80-9204-42a56d4dde89-kube-api-access-lwpb5\") pod \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.840941 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-config-data\") pod \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\" (UID: \"9ee77a28-a1b7-4d80-9204-42a56d4dde89\") " Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.845862 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-scripts" (OuterVolumeSpecName: "scripts") pod "9ee77a28-a1b7-4d80-9204-42a56d4dde89" (UID: "9ee77a28-a1b7-4d80-9204-42a56d4dde89"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.845912 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ee77a28-a1b7-4d80-9204-42a56d4dde89-kube-api-access-lwpb5" (OuterVolumeSpecName: "kube-api-access-lwpb5") pod "9ee77a28-a1b7-4d80-9204-42a56d4dde89" (UID: "9ee77a28-a1b7-4d80-9204-42a56d4dde89"). InnerVolumeSpecName "kube-api-access-lwpb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.886878 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-config-data" (OuterVolumeSpecName: "config-data") pod "9ee77a28-a1b7-4d80-9204-42a56d4dde89" (UID: "9ee77a28-a1b7-4d80-9204-42a56d4dde89"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.887664 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ee77a28-a1b7-4d80-9204-42a56d4dde89" (UID: "9ee77a28-a1b7-4d80-9204-42a56d4dde89"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.942918 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.943274 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.943287 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwpb5\" (UniqueName: \"kubernetes.io/projected/9ee77a28-a1b7-4d80-9204-42a56d4dde89-kube-api-access-lwpb5\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.943301 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ee77a28-a1b7-4d80-9204-42a56d4dde89-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.972708 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.972763 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:19:40 crc kubenswrapper[4726]: I1001 16:19:40.997996 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.056364 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.062391 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-czbjh"] Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.062838 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-czbjh" podUID="d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" containerName="dnsmasq-dns" containerID="cri-o://b3856e2da018ee9d1b6a1b9c847e0df283e574a77b4f04d7320b96e91a935ce7" gracePeriod=10 Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.485956 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qggxs" event={"ID":"9ee77a28-a1b7-4d80-9204-42a56d4dde89","Type":"ContainerDied","Data":"f0cf93ab7aee5a7f7343f73503492ce0edfe145e526e632a1c33c343d0a5bd87"} Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.486249 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0cf93ab7aee5a7f7343f73503492ce0edfe145e526e632a1c33c343d0a5bd87" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.486316 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qggxs" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.511771 4726 generic.go:334] "Generic (PLEG): container finished" podID="d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" containerID="b3856e2da018ee9d1b6a1b9c847e0df283e574a77b4f04d7320b96e91a935ce7" exitCode=0 Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.511862 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-czbjh" event={"ID":"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e","Type":"ContainerDied","Data":"b3856e2da018ee9d1b6a1b9c847e0df283e574a77b4f04d7320b96e91a935ce7"} Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.538317 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9906cce7-2c94-4290-971e-d8a556a3a745","Type":"ContainerStarted","Data":"9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e"} Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.538361 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9906cce7-2c94-4290-971e-d8a556a3a745","Type":"ContainerStarted","Data":"1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710"} Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.586011 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.585989672 podStartE2EDuration="2.585989672s" podCreationTimestamp="2025-10-01 16:19:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:19:41.568748856 +0000 UTC m=+1234.470301443" watchObservedRunningTime="2025-10-01 16:19:41.585989672 +0000 UTC m=+1234.487542249" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.665897 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.690127 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.698619 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.698851 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="55ae01da-8b6c-464b-b77b-5f0358fd8b18" containerName="nova-api-log" containerID="cri-o://e96fa9dba30321a23befb6390facdced1fd809bfae6624f8d38080a92a72d9d6" gracePeriod=30 Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.698996 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="55ae01da-8b6c-464b-b77b-5f0358fd8b18" containerName="nova-api-api" containerID="cri-o://fec82a33896f952be8d9a59bf54762dc24e72dca6807e461b2eca2d18efb2581" gracePeriod=30 Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.712338 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="55ae01da-8b6c-464b-b77b-5f0358fd8b18" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.188:8774/\": EOF" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.712543 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="55ae01da-8b6c-464b-b77b-5f0358fd8b18" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.188:8774/\": EOF" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.724191 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.735533 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.773927 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-ovsdbserver-nb\") pod \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.773985 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-config\") pod \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.774183 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-ovsdbserver-sb\") pod \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.774227 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-dns-swift-storage-0\") pod \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.774267 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qf7r\" (UniqueName: \"kubernetes.io/projected/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-kube-api-access-2qf7r\") pod \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.774344 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-dns-svc\") pod \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\" (UID: \"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e\") " Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.797725 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-kube-api-access-2qf7r" (OuterVolumeSpecName: "kube-api-access-2qf7r") pod "d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" (UID: "d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e"). InnerVolumeSpecName "kube-api-access-2qf7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.876698 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qf7r\" (UniqueName: \"kubernetes.io/projected/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-kube-api-access-2qf7r\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.883810 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" (UID: "d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.884497 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-config" (OuterVolumeSpecName: "config") pod "d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" (UID: "d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.898511 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" (UID: "d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.912686 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" (UID: "d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.924513 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" (UID: "d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.978361 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.978401 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.978415 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.978426 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:41 crc kubenswrapper[4726]: I1001 16:19:41.978438 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:42 crc kubenswrapper[4726]: I1001 16:19:42.545479 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-czbjh" event={"ID":"d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e","Type":"ContainerDied","Data":"dd395c66044a73144b33efab265ad3bfc8f1c2fe11e20bdd847657ae2b3307ea"} Oct 01 16:19:42 crc kubenswrapper[4726]: I1001 16:19:42.545523 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-czbjh" Oct 01 16:19:42 crc kubenswrapper[4726]: I1001 16:19:42.545540 4726 scope.go:117] "RemoveContainer" containerID="b3856e2da018ee9d1b6a1b9c847e0df283e574a77b4f04d7320b96e91a935ce7" Oct 01 16:19:42 crc kubenswrapper[4726]: I1001 16:19:42.547598 4726 generic.go:334] "Generic (PLEG): container finished" podID="55ae01da-8b6c-464b-b77b-5f0358fd8b18" containerID="e96fa9dba30321a23befb6390facdced1fd809bfae6624f8d38080a92a72d9d6" exitCode=143 Oct 01 16:19:42 crc kubenswrapper[4726]: I1001 16:19:42.547658 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55ae01da-8b6c-464b-b77b-5f0358fd8b18","Type":"ContainerDied","Data":"e96fa9dba30321a23befb6390facdced1fd809bfae6624f8d38080a92a72d9d6"} Oct 01 16:19:42 crc kubenswrapper[4726]: I1001 16:19:42.554254 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fdf946f7-e990-4849-a998-3d151a619bf2","Type":"ContainerStarted","Data":"ed56207bdcf0dc3ac3e80204eacc9ccaadc4494bafb39c317cd690d9e298a0a0"} Oct 01 16:19:42 crc kubenswrapper[4726]: I1001 16:19:42.554804 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 16:19:42 crc kubenswrapper[4726]: I1001 16:19:42.567639 4726 scope.go:117] "RemoveContainer" containerID="671c329f9be6ecf907b98714e983c08724dc01956ed7e1321a43622a60724b1c" Oct 01 16:19:42 crc kubenswrapper[4726]: I1001 16:19:42.600923 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.479122921 podStartE2EDuration="9.600902916s" podCreationTimestamp="2025-10-01 16:19:33 +0000 UTC" firstStartedPulling="2025-10-01 16:19:37.637490851 +0000 UTC m=+1230.539043428" lastFinishedPulling="2025-10-01 16:19:41.759270846 +0000 UTC m=+1234.660823423" observedRunningTime="2025-10-01 16:19:42.599896038 +0000 UTC m=+1235.501448615" watchObservedRunningTime="2025-10-01 16:19:42.600902916 +0000 UTC m=+1235.502455493" Oct 01 16:19:42 crc kubenswrapper[4726]: I1001 16:19:42.635385 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-czbjh"] Oct 01 16:19:42 crc kubenswrapper[4726]: I1001 16:19:42.643735 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-czbjh"] Oct 01 16:19:43 crc kubenswrapper[4726]: I1001 16:19:43.578289 4726 generic.go:334] "Generic (PLEG): container finished" podID="d747a1f6-2d98-4b9e-9a0d-833d0afede9d" containerID="ed81401bbb74ca249f20375c440e14f84f09d3a54734aef9dc6aed10e691944d" exitCode=0 Oct 01 16:19:43 crc kubenswrapper[4726]: I1001 16:19:43.578377 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-79vp9" event={"ID":"d747a1f6-2d98-4b9e-9a0d-833d0afede9d","Type":"ContainerDied","Data":"ed81401bbb74ca249f20375c440e14f84f09d3a54734aef9dc6aed10e691944d"} Oct 01 16:19:43 crc kubenswrapper[4726]: I1001 16:19:43.582377 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6" containerName="nova-scheduler-scheduler" containerID="cri-o://dcb394db92f47b55646f5397129b6f8bcf575fe56c9d6b29dc7050e0cb1a56fc" gracePeriod=30 Oct 01 16:19:43 crc kubenswrapper[4726]: I1001 16:19:43.582573 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9906cce7-2c94-4290-971e-d8a556a3a745" containerName="nova-metadata-log" containerID="cri-o://1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710" gracePeriod=30 Oct 01 16:19:43 crc kubenswrapper[4726]: I1001 16:19:43.582631 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9906cce7-2c94-4290-971e-d8a556a3a745" containerName="nova-metadata-metadata" containerID="cri-o://9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e" gracePeriod=30 Oct 01 16:19:43 crc kubenswrapper[4726]: I1001 16:19:43.832323 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" path="/var/lib/kubelet/pods/d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e/volumes" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.203801 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.322747 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-combined-ca-bundle\") pod \"9906cce7-2c94-4290-971e-d8a556a3a745\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.322909 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9906cce7-2c94-4290-971e-d8a556a3a745-logs\") pod \"9906cce7-2c94-4290-971e-d8a556a3a745\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.322947 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9zk4\" (UniqueName: \"kubernetes.io/projected/9906cce7-2c94-4290-971e-d8a556a3a745-kube-api-access-s9zk4\") pod \"9906cce7-2c94-4290-971e-d8a556a3a745\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.323102 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-config-data\") pod \"9906cce7-2c94-4290-971e-d8a556a3a745\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.323164 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-nova-metadata-tls-certs\") pod \"9906cce7-2c94-4290-971e-d8a556a3a745\" (UID: \"9906cce7-2c94-4290-971e-d8a556a3a745\") " Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.323388 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9906cce7-2c94-4290-971e-d8a556a3a745-logs" (OuterVolumeSpecName: "logs") pod "9906cce7-2c94-4290-971e-d8a556a3a745" (UID: "9906cce7-2c94-4290-971e-d8a556a3a745"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.323932 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9906cce7-2c94-4290-971e-d8a556a3a745-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.356431 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9906cce7-2c94-4290-971e-d8a556a3a745-kube-api-access-s9zk4" (OuterVolumeSpecName: "kube-api-access-s9zk4") pod "9906cce7-2c94-4290-971e-d8a556a3a745" (UID: "9906cce7-2c94-4290-971e-d8a556a3a745"). InnerVolumeSpecName "kube-api-access-s9zk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.369018 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9906cce7-2c94-4290-971e-d8a556a3a745" (UID: "9906cce7-2c94-4290-971e-d8a556a3a745"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.375535 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-config-data" (OuterVolumeSpecName: "config-data") pod "9906cce7-2c94-4290-971e-d8a556a3a745" (UID: "9906cce7-2c94-4290-971e-d8a556a3a745"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.385697 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "9906cce7-2c94-4290-971e-d8a556a3a745" (UID: "9906cce7-2c94-4290-971e-d8a556a3a745"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.425573 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.425619 4726 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.425628 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9906cce7-2c94-4290-971e-d8a556a3a745-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.425637 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9zk4\" (UniqueName: \"kubernetes.io/projected/9906cce7-2c94-4290-971e-d8a556a3a745-kube-api-access-s9zk4\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.591059 4726 generic.go:334] "Generic (PLEG): container finished" podID="9906cce7-2c94-4290-971e-d8a556a3a745" containerID="9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e" exitCode=0 Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.591097 4726 generic.go:334] "Generic (PLEG): container finished" podID="9906cce7-2c94-4290-971e-d8a556a3a745" containerID="1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710" exitCode=143 Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.591147 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9906cce7-2c94-4290-971e-d8a556a3a745","Type":"ContainerDied","Data":"9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e"} Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.591192 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9906cce7-2c94-4290-971e-d8a556a3a745","Type":"ContainerDied","Data":"1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710"} Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.591206 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9906cce7-2c94-4290-971e-d8a556a3a745","Type":"ContainerDied","Data":"9eba22707533e9fc27488ac5b6b69640ab3dcd0d9d3ba29e039bf417407bc181"} Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.591225 4726 scope.go:117] "RemoveContainer" containerID="9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.591148 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.621373 4726 scope.go:117] "RemoveContainer" containerID="1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.639853 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.649754 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.663759 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:44 crc kubenswrapper[4726]: E1001 16:19:44.664167 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9906cce7-2c94-4290-971e-d8a556a3a745" containerName="nova-metadata-log" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.664187 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9906cce7-2c94-4290-971e-d8a556a3a745" containerName="nova-metadata-log" Oct 01 16:19:44 crc kubenswrapper[4726]: E1001 16:19:44.664198 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9906cce7-2c94-4290-971e-d8a556a3a745" containerName="nova-metadata-metadata" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.664205 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9906cce7-2c94-4290-971e-d8a556a3a745" containerName="nova-metadata-metadata" Oct 01 16:19:44 crc kubenswrapper[4726]: E1001 16:19:44.664214 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" containerName="init" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.664220 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" containerName="init" Oct 01 16:19:44 crc kubenswrapper[4726]: E1001 16:19:44.664233 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ee77a28-a1b7-4d80-9204-42a56d4dde89" containerName="nova-manage" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.664240 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ee77a28-a1b7-4d80-9204-42a56d4dde89" containerName="nova-manage" Oct 01 16:19:44 crc kubenswrapper[4726]: E1001 16:19:44.664262 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" containerName="dnsmasq-dns" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.664267 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" containerName="dnsmasq-dns" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.664459 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9906cce7-2c94-4290-971e-d8a556a3a745" containerName="nova-metadata-log" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.664478 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0810a3d-a5f7-4c7a-bc64-8b46e7eab64e" containerName="dnsmasq-dns" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.664497 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ee77a28-a1b7-4d80-9204-42a56d4dde89" containerName="nova-manage" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.664521 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9906cce7-2c94-4290-971e-d8a556a3a745" containerName="nova-metadata-metadata" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.665536 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.668232 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.669082 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.671359 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.675409 4726 scope.go:117] "RemoveContainer" containerID="9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e" Oct 01 16:19:44 crc kubenswrapper[4726]: E1001 16:19:44.675806 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e\": container with ID starting with 9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e not found: ID does not exist" containerID="9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.675848 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e"} err="failed to get container status \"9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e\": rpc error: code = NotFound desc = could not find container \"9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e\": container with ID starting with 9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e not found: ID does not exist" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.675873 4726 scope.go:117] "RemoveContainer" containerID="1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710" Oct 01 16:19:44 crc kubenswrapper[4726]: E1001 16:19:44.676110 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710\": container with ID starting with 1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710 not found: ID does not exist" containerID="1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.676132 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710"} err="failed to get container status \"1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710\": rpc error: code = NotFound desc = could not find container \"1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710\": container with ID starting with 1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710 not found: ID does not exist" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.676146 4726 scope.go:117] "RemoveContainer" containerID="9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.676302 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e"} err="failed to get container status \"9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e\": rpc error: code = NotFound desc = could not find container \"9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e\": container with ID starting with 9b1f5ba140ca3368bcbacfec8a59a08d071aab9df634956be9e8dedbc2b9724e not found: ID does not exist" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.676329 4726 scope.go:117] "RemoveContainer" containerID="1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.676488 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710"} err="failed to get container status \"1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710\": rpc error: code = NotFound desc = could not find container \"1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710\": container with ID starting with 1554f64c0b44f231fa1c9fe54fb9a6f17eb8bffa3662f98c63b60f50e8dde710 not found: ID does not exist" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.731029 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnmn5\" (UniqueName: \"kubernetes.io/projected/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-kube-api-access-nnmn5\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.731094 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.731168 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-logs\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.731199 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-config-data\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.731261 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.833096 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-config-data\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.833198 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.833261 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnmn5\" (UniqueName: \"kubernetes.io/projected/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-kube-api-access-nnmn5\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.833282 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.833980 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-logs\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.834769 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-logs\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.838775 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.839618 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-config-data\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.848727 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:44 crc kubenswrapper[4726]: I1001 16:19:44.857980 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnmn5\" (UniqueName: \"kubernetes.io/projected/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-kube-api-access-nnmn5\") pod \"nova-metadata-0\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " pod="openstack/nova-metadata-0" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.035382 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.038183 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-combined-ca-bundle\") pod \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.038359 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9bfs\" (UniqueName: \"kubernetes.io/projected/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-kube-api-access-s9bfs\") pod \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.038463 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-config-data\") pod \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.038520 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-scripts\") pod \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\" (UID: \"d747a1f6-2d98-4b9e-9a0d-833d0afede9d\") " Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.049735 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-kube-api-access-s9bfs" (OuterVolumeSpecName: "kube-api-access-s9bfs") pod "d747a1f6-2d98-4b9e-9a0d-833d0afede9d" (UID: "d747a1f6-2d98-4b9e-9a0d-833d0afede9d"). InnerVolumeSpecName "kube-api-access-s9bfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.050279 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.051287 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-scripts" (OuterVolumeSpecName: "scripts") pod "d747a1f6-2d98-4b9e-9a0d-833d0afede9d" (UID: "d747a1f6-2d98-4b9e-9a0d-833d0afede9d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.087769 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d747a1f6-2d98-4b9e-9a0d-833d0afede9d" (UID: "d747a1f6-2d98-4b9e-9a0d-833d0afede9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.095565 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-config-data" (OuterVolumeSpecName: "config-data") pod "d747a1f6-2d98-4b9e-9a0d-833d0afede9d" (UID: "d747a1f6-2d98-4b9e-9a0d-833d0afede9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.140172 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9bfs\" (UniqueName: \"kubernetes.io/projected/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-kube-api-access-s9bfs\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.140381 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.140393 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.140403 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d747a1f6-2d98-4b9e-9a0d-833d0afede9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.566423 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.601816 4726 generic.go:334] "Generic (PLEG): container finished" podID="3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6" containerID="dcb394db92f47b55646f5397129b6f8bcf575fe56c9d6b29dc7050e0cb1a56fc" exitCode=0 Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.601863 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6","Type":"ContainerDied","Data":"dcb394db92f47b55646f5397129b6f8bcf575fe56c9d6b29dc7050e0cb1a56fc"} Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.613146 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56e4e1ae-a938-4cf3-a34b-654dbcebeb51","Type":"ContainerStarted","Data":"e23d6ff9abec3c388c3fe07f77eb01ee3aece7f51ffa7d70ca929256fd2dfc14"} Oct 01 16:19:45 crc kubenswrapper[4726]: E1001 16:19:45.613736 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of dcb394db92f47b55646f5397129b6f8bcf575fe56c9d6b29dc7050e0cb1a56fc is running failed: container process not found" containerID="dcb394db92f47b55646f5397129b6f8bcf575fe56c9d6b29dc7050e0cb1a56fc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 16:19:45 crc kubenswrapper[4726]: E1001 16:19:45.619376 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of dcb394db92f47b55646f5397129b6f8bcf575fe56c9d6b29dc7050e0cb1a56fc is running failed: container process not found" containerID="dcb394db92f47b55646f5397129b6f8bcf575fe56c9d6b29dc7050e0cb1a56fc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 16:19:45 crc kubenswrapper[4726]: E1001 16:19:45.619687 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of dcb394db92f47b55646f5397129b6f8bcf575fe56c9d6b29dc7050e0cb1a56fc is running failed: container process not found" containerID="dcb394db92f47b55646f5397129b6f8bcf575fe56c9d6b29dc7050e0cb1a56fc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 16:19:45 crc kubenswrapper[4726]: E1001 16:19:45.619716 4726 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of dcb394db92f47b55646f5397129b6f8bcf575fe56c9d6b29dc7050e0cb1a56fc is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6" containerName="nova-scheduler-scheduler" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.620790 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-79vp9" event={"ID":"d747a1f6-2d98-4b9e-9a0d-833d0afede9d","Type":"ContainerDied","Data":"9c7c00e6c80b563fd49bc8e33e4022b93e4376d6a9709d2b1ed0621acc9e8c58"} Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.620818 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c7c00e6c80b563fd49bc8e33e4022b93e4376d6a9709d2b1ed0621acc9e8c58" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.620879 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-79vp9" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.680235 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 16:19:45 crc kubenswrapper[4726]: E1001 16:19:45.680820 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d747a1f6-2d98-4b9e-9a0d-833d0afede9d" containerName="nova-cell1-conductor-db-sync" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.680846 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d747a1f6-2d98-4b9e-9a0d-833d0afede9d" containerName="nova-cell1-conductor-db-sync" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.681028 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d747a1f6-2d98-4b9e-9a0d-833d0afede9d" containerName="nova-cell1-conductor-db-sync" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.681880 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.682930 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.684502 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.689348 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.853325 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9906cce7-2c94-4290-971e-d8a556a3a745" path="/var/lib/kubelet/pods/9906cce7-2c94-4290-971e-d8a556a3a745/volumes" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.855107 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66g59\" (UniqueName: \"kubernetes.io/projected/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-kube-api-access-66g59\") pod \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\" (UID: \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\") " Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.855273 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-config-data\") pod \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\" (UID: \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\") " Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.855393 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-combined-ca-bundle\") pod \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\" (UID: \"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6\") " Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.855728 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/812fbec4-8c87-4e63-b7e6-e165412dc519-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"812fbec4-8c87-4e63-b7e6-e165412dc519\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.855817 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84jbn\" (UniqueName: \"kubernetes.io/projected/812fbec4-8c87-4e63-b7e6-e165412dc519-kube-api-access-84jbn\") pod \"nova-cell1-conductor-0\" (UID: \"812fbec4-8c87-4e63-b7e6-e165412dc519\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.856056 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/812fbec4-8c87-4e63-b7e6-e165412dc519-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"812fbec4-8c87-4e63-b7e6-e165412dc519\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.864369 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-kube-api-access-66g59" (OuterVolumeSpecName: "kube-api-access-66g59") pod "3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6" (UID: "3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6"). InnerVolumeSpecName "kube-api-access-66g59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.900183 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6" (UID: "3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.912249 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-config-data" (OuterVolumeSpecName: "config-data") pod "3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6" (UID: "3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.957957 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/812fbec4-8c87-4e63-b7e6-e165412dc519-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"812fbec4-8c87-4e63-b7e6-e165412dc519\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.958052 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/812fbec4-8c87-4e63-b7e6-e165412dc519-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"812fbec4-8c87-4e63-b7e6-e165412dc519\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.958107 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84jbn\" (UniqueName: \"kubernetes.io/projected/812fbec4-8c87-4e63-b7e6-e165412dc519-kube-api-access-84jbn\") pod \"nova-cell1-conductor-0\" (UID: \"812fbec4-8c87-4e63-b7e6-e165412dc519\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.958274 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66g59\" (UniqueName: \"kubernetes.io/projected/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-kube-api-access-66g59\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.958293 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.958307 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.963825 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/812fbec4-8c87-4e63-b7e6-e165412dc519-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"812fbec4-8c87-4e63-b7e6-e165412dc519\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.977707 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/812fbec4-8c87-4e63-b7e6-e165412dc519-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"812fbec4-8c87-4e63-b7e6-e165412dc519\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:19:45 crc kubenswrapper[4726]: I1001 16:19:45.981166 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84jbn\" (UniqueName: \"kubernetes.io/projected/812fbec4-8c87-4e63-b7e6-e165412dc519-kube-api-access-84jbn\") pod \"nova-cell1-conductor-0\" (UID: \"812fbec4-8c87-4e63-b7e6-e165412dc519\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.086319 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.555199 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.636880 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6","Type":"ContainerDied","Data":"b50cb0626bb1f0aac79c3b8bc58e97c88ed1975a21790d2c9c3f53b05e61e2d9"} Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.636912 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.636946 4726 scope.go:117] "RemoveContainer" containerID="dcb394db92f47b55646f5397129b6f8bcf575fe56c9d6b29dc7050e0cb1a56fc" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.637705 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"812fbec4-8c87-4e63-b7e6-e165412dc519","Type":"ContainerStarted","Data":"32c0984d6e09f92c24a5daceb716fd47ab0ee52209a91cbd791bfa2b02506cce"} Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.639772 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56e4e1ae-a938-4cf3-a34b-654dbcebeb51","Type":"ContainerStarted","Data":"bd147ac73eff6c440ece5ffed7152fa1c5387f469aaf8b342ffd9a6cbc1b294e"} Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.639809 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56e4e1ae-a938-4cf3-a34b-654dbcebeb51","Type":"ContainerStarted","Data":"d902b6ac6bfcf6d4a6342d6c60c1495659de816e5933d0a107c9dbf9af46fc98"} Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.664887 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.664871201 podStartE2EDuration="2.664871201s" podCreationTimestamp="2025-10-01 16:19:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:19:46.657681219 +0000 UTC m=+1239.559233816" watchObservedRunningTime="2025-10-01 16:19:46.664871201 +0000 UTC m=+1239.566423778" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.708846 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.716727 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.731737 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:19:46 crc kubenswrapper[4726]: E1001 16:19:46.732174 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6" containerName="nova-scheduler-scheduler" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.732194 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6" containerName="nova-scheduler-scheduler" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.732414 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6" containerName="nova-scheduler-scheduler" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.733141 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.735906 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.741580 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.874914 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gbrd\" (UniqueName: \"kubernetes.io/projected/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-kube-api-access-7gbrd\") pod \"nova-scheduler-0\" (UID: \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.875231 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-config-data\") pod \"nova-scheduler-0\" (UID: \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.875325 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.976789 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.976931 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gbrd\" (UniqueName: \"kubernetes.io/projected/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-kube-api-access-7gbrd\") pod \"nova-scheduler-0\" (UID: \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.976968 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-config-data\") pod \"nova-scheduler-0\" (UID: \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.983371 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-config-data\") pod \"nova-scheduler-0\" (UID: \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.986751 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:46 crc kubenswrapper[4726]: I1001 16:19:46.996788 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gbrd\" (UniqueName: \"kubernetes.io/projected/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-kube-api-access-7gbrd\") pod \"nova-scheduler-0\" (UID: \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\") " pod="openstack/nova-scheduler-0" Oct 01 16:19:47 crc kubenswrapper[4726]: I1001 16:19:47.078494 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:19:47 crc kubenswrapper[4726]: I1001 16:19:47.516632 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:19:47 crc kubenswrapper[4726]: W1001 16:19:47.518087 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d6747b1_097c_4c78_ad2f_1f2575dd5e46.slice/crio-f79afda308a18c7b4caeace34a93afd6e5f005d4b8e869b9731bd378b660a29b WatchSource:0}: Error finding container f79afda308a18c7b4caeace34a93afd6e5f005d4b8e869b9731bd378b660a29b: Status 404 returned error can't find the container with id f79afda308a18c7b4caeace34a93afd6e5f005d4b8e869b9731bd378b660a29b Oct 01 16:19:47 crc kubenswrapper[4726]: I1001 16:19:47.668227 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7d6747b1-097c-4c78-ad2f-1f2575dd5e46","Type":"ContainerStarted","Data":"f79afda308a18c7b4caeace34a93afd6e5f005d4b8e869b9731bd378b660a29b"} Oct 01 16:19:47 crc kubenswrapper[4726]: I1001 16:19:47.672902 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"812fbec4-8c87-4e63-b7e6-e165412dc519","Type":"ContainerStarted","Data":"be0f71f7c84f0ce2f35935e22443d0d2f99b287f02f029e9df5a25953c824bb6"} Oct 01 16:19:47 crc kubenswrapper[4726]: I1001 16:19:47.672996 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 01 16:19:47 crc kubenswrapper[4726]: I1001 16:19:47.695813 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.695796646 podStartE2EDuration="2.695796646s" podCreationTimestamp="2025-10-01 16:19:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:19:47.689694714 +0000 UTC m=+1240.591247291" watchObservedRunningTime="2025-10-01 16:19:47.695796646 +0000 UTC m=+1240.597349223" Oct 01 16:19:47 crc kubenswrapper[4726]: I1001 16:19:47.823001 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6" path="/var/lib/kubelet/pods/3ca4c81f-f0da-4813-8ad6-3824ba1ce8e6/volumes" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.647972 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.683565 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7d6747b1-097c-4c78-ad2f-1f2575dd5e46","Type":"ContainerStarted","Data":"e21317068070fadc0b29bd39c43364f8a1cdeaa244507389dbbf6e4f20858276"} Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.689382 4726 generic.go:334] "Generic (PLEG): container finished" podID="55ae01da-8b6c-464b-b77b-5f0358fd8b18" containerID="fec82a33896f952be8d9a59bf54762dc24e72dca6807e461b2eca2d18efb2581" exitCode=0 Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.690332 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.690527 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55ae01da-8b6c-464b-b77b-5f0358fd8b18","Type":"ContainerDied","Data":"fec82a33896f952be8d9a59bf54762dc24e72dca6807e461b2eca2d18efb2581"} Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.690562 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55ae01da-8b6c-464b-b77b-5f0358fd8b18","Type":"ContainerDied","Data":"e52d91953a8f2dcf15e046bac96c3a65f74de10adbfc64e1e6583954961a08ee"} Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.690583 4726 scope.go:117] "RemoveContainer" containerID="fec82a33896f952be8d9a59bf54762dc24e72dca6807e461b2eca2d18efb2581" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.708518 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.7084942869999997 podStartE2EDuration="2.708494287s" podCreationTimestamp="2025-10-01 16:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:19:48.696487348 +0000 UTC m=+1241.598039935" watchObservedRunningTime="2025-10-01 16:19:48.708494287 +0000 UTC m=+1241.610046864" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.737249 4726 scope.go:117] "RemoveContainer" containerID="e96fa9dba30321a23befb6390facdced1fd809bfae6624f8d38080a92a72d9d6" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.761663 4726 scope.go:117] "RemoveContainer" containerID="fec82a33896f952be8d9a59bf54762dc24e72dca6807e461b2eca2d18efb2581" Oct 01 16:19:48 crc kubenswrapper[4726]: E1001 16:19:48.762365 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fec82a33896f952be8d9a59bf54762dc24e72dca6807e461b2eca2d18efb2581\": container with ID starting with fec82a33896f952be8d9a59bf54762dc24e72dca6807e461b2eca2d18efb2581 not found: ID does not exist" containerID="fec82a33896f952be8d9a59bf54762dc24e72dca6807e461b2eca2d18efb2581" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.762442 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fec82a33896f952be8d9a59bf54762dc24e72dca6807e461b2eca2d18efb2581"} err="failed to get container status \"fec82a33896f952be8d9a59bf54762dc24e72dca6807e461b2eca2d18efb2581\": rpc error: code = NotFound desc = could not find container \"fec82a33896f952be8d9a59bf54762dc24e72dca6807e461b2eca2d18efb2581\": container with ID starting with fec82a33896f952be8d9a59bf54762dc24e72dca6807e461b2eca2d18efb2581 not found: ID does not exist" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.762520 4726 scope.go:117] "RemoveContainer" containerID="e96fa9dba30321a23befb6390facdced1fd809bfae6624f8d38080a92a72d9d6" Oct 01 16:19:48 crc kubenswrapper[4726]: E1001 16:19:48.762958 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e96fa9dba30321a23befb6390facdced1fd809bfae6624f8d38080a92a72d9d6\": container with ID starting with e96fa9dba30321a23befb6390facdced1fd809bfae6624f8d38080a92a72d9d6 not found: ID does not exist" containerID="e96fa9dba30321a23befb6390facdced1fd809bfae6624f8d38080a92a72d9d6" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.762991 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e96fa9dba30321a23befb6390facdced1fd809bfae6624f8d38080a92a72d9d6"} err="failed to get container status \"e96fa9dba30321a23befb6390facdced1fd809bfae6624f8d38080a92a72d9d6\": rpc error: code = NotFound desc = could not find container \"e96fa9dba30321a23befb6390facdced1fd809bfae6624f8d38080a92a72d9d6\": container with ID starting with e96fa9dba30321a23befb6390facdced1fd809bfae6624f8d38080a92a72d9d6 not found: ID does not exist" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.810003 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ae01da-8b6c-464b-b77b-5f0358fd8b18-combined-ca-bundle\") pod \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.810105 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55ae01da-8b6c-464b-b77b-5f0358fd8b18-logs\") pod \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.810148 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ae01da-8b6c-464b-b77b-5f0358fd8b18-config-data\") pod \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.810214 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjx2r\" (UniqueName: \"kubernetes.io/projected/55ae01da-8b6c-464b-b77b-5f0358fd8b18-kube-api-access-rjx2r\") pod \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\" (UID: \"55ae01da-8b6c-464b-b77b-5f0358fd8b18\") " Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.812226 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55ae01da-8b6c-464b-b77b-5f0358fd8b18-logs" (OuterVolumeSpecName: "logs") pod "55ae01da-8b6c-464b-b77b-5f0358fd8b18" (UID: "55ae01da-8b6c-464b-b77b-5f0358fd8b18"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.821942 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55ae01da-8b6c-464b-b77b-5f0358fd8b18-kube-api-access-rjx2r" (OuterVolumeSpecName: "kube-api-access-rjx2r") pod "55ae01da-8b6c-464b-b77b-5f0358fd8b18" (UID: "55ae01da-8b6c-464b-b77b-5f0358fd8b18"). InnerVolumeSpecName "kube-api-access-rjx2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.841154 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55ae01da-8b6c-464b-b77b-5f0358fd8b18-config-data" (OuterVolumeSpecName: "config-data") pod "55ae01da-8b6c-464b-b77b-5f0358fd8b18" (UID: "55ae01da-8b6c-464b-b77b-5f0358fd8b18"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.850485 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55ae01da-8b6c-464b-b77b-5f0358fd8b18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55ae01da-8b6c-464b-b77b-5f0358fd8b18" (UID: "55ae01da-8b6c-464b-b77b-5f0358fd8b18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.913622 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ae01da-8b6c-464b-b77b-5f0358fd8b18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.913670 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55ae01da-8b6c-464b-b77b-5f0358fd8b18-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.913681 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ae01da-8b6c-464b-b77b-5f0358fd8b18-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:48 crc kubenswrapper[4726]: I1001 16:19:48.913689 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjx2r\" (UniqueName: \"kubernetes.io/projected/55ae01da-8b6c-464b-b77b-5f0358fd8b18-kube-api-access-rjx2r\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.029385 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.044130 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.057503 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 16:19:49 crc kubenswrapper[4726]: E1001 16:19:49.058024 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55ae01da-8b6c-464b-b77b-5f0358fd8b18" containerName="nova-api-log" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.058044 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="55ae01da-8b6c-464b-b77b-5f0358fd8b18" containerName="nova-api-log" Oct 01 16:19:49 crc kubenswrapper[4726]: E1001 16:19:49.058085 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55ae01da-8b6c-464b-b77b-5f0358fd8b18" containerName="nova-api-api" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.058095 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="55ae01da-8b6c-464b-b77b-5f0358fd8b18" containerName="nova-api-api" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.058303 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="55ae01da-8b6c-464b-b77b-5f0358fd8b18" containerName="nova-api-api" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.058330 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="55ae01da-8b6c-464b-b77b-5f0358fd8b18" containerName="nova-api-log" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.059283 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.061424 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.072484 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.221218 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phsnk\" (UniqueName: \"kubernetes.io/projected/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-kube-api-access-phsnk\") pod \"nova-api-0\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " pod="openstack/nova-api-0" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.221260 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-logs\") pod \"nova-api-0\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " pod="openstack/nova-api-0" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.221289 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-config-data\") pod \"nova-api-0\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " pod="openstack/nova-api-0" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.221372 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " pod="openstack/nova-api-0" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.323608 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " pod="openstack/nova-api-0" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.323841 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phsnk\" (UniqueName: \"kubernetes.io/projected/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-kube-api-access-phsnk\") pod \"nova-api-0\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " pod="openstack/nova-api-0" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.323893 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-logs\") pod \"nova-api-0\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " pod="openstack/nova-api-0" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.323946 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-config-data\") pod \"nova-api-0\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " pod="openstack/nova-api-0" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.324624 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-logs\") pod \"nova-api-0\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " pod="openstack/nova-api-0" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.329767 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-config-data\") pod \"nova-api-0\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " pod="openstack/nova-api-0" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.329768 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " pod="openstack/nova-api-0" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.376820 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phsnk\" (UniqueName: \"kubernetes.io/projected/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-kube-api-access-phsnk\") pod \"nova-api-0\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " pod="openstack/nova-api-0" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.383833 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.831946 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55ae01da-8b6c-464b-b77b-5f0358fd8b18" path="/var/lib/kubelet/pods/55ae01da-8b6c-464b-b77b-5f0358fd8b18/volumes" Oct 01 16:19:49 crc kubenswrapper[4726]: I1001 16:19:49.964856 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:19:50 crc kubenswrapper[4726]: I1001 16:19:50.051234 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 16:19:50 crc kubenswrapper[4726]: I1001 16:19:50.051497 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 16:19:50 crc kubenswrapper[4726]: I1001 16:19:50.708457 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76","Type":"ContainerStarted","Data":"9698c8bd47f1703eb9e267cfc613890b07bf7545abe58ab8c41065f3e6e9c86e"} Oct 01 16:19:50 crc kubenswrapper[4726]: I1001 16:19:50.708535 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76","Type":"ContainerStarted","Data":"fcb12de2fb93ae3828d3e48cc8c8cde46249cf78784ab46772f8d1650f7a86bb"} Oct 01 16:19:50 crc kubenswrapper[4726]: I1001 16:19:50.708550 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76","Type":"ContainerStarted","Data":"0e33d9ef7fa6bf6e792497953f4fb080826deffb77b0ee3ec3d60ecec6272bd7"} Oct 01 16:19:50 crc kubenswrapper[4726]: I1001 16:19:50.737016 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.736998646 podStartE2EDuration="1.736998646s" podCreationTimestamp="2025-10-01 16:19:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:19:50.730900495 +0000 UTC m=+1243.632453082" watchObservedRunningTime="2025-10-01 16:19:50.736998646 +0000 UTC m=+1243.638551223" Oct 01 16:19:51 crc kubenswrapper[4726]: I1001 16:19:51.115479 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 01 16:19:52 crc kubenswrapper[4726]: I1001 16:19:52.079526 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 16:19:55 crc kubenswrapper[4726]: I1001 16:19:55.051349 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 16:19:55 crc kubenswrapper[4726]: I1001 16:19:55.051879 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 16:19:56 crc kubenswrapper[4726]: I1001 16:19:56.067267 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:19:56 crc kubenswrapper[4726]: I1001 16:19:56.067275 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:19:57 crc kubenswrapper[4726]: I1001 16:19:57.079176 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 01 16:19:57 crc kubenswrapper[4726]: I1001 16:19:57.104274 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 01 16:19:57 crc kubenswrapper[4726]: I1001 16:19:57.803290 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 01 16:19:59 crc kubenswrapper[4726]: I1001 16:19:59.385131 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:19:59 crc kubenswrapper[4726]: I1001 16:19:59.385425 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:20:00 crc kubenswrapper[4726]: I1001 16:20:00.467280 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 16:20:00 crc kubenswrapper[4726]: I1001 16:20:00.467313 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 16:20:03 crc kubenswrapper[4726]: I1001 16:20:03.817914 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 01 16:20:05 crc kubenswrapper[4726]: I1001 16:20:05.056005 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 16:20:05 crc kubenswrapper[4726]: I1001 16:20:05.057400 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 16:20:05 crc kubenswrapper[4726]: I1001 16:20:05.061469 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 16:20:05 crc kubenswrapper[4726]: I1001 16:20:05.905390 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 16:20:08 crc kubenswrapper[4726]: I1001 16:20:08.867071 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:08 crc kubenswrapper[4726]: I1001 16:20:08.926097 4726 generic.go:334] "Generic (PLEG): container finished" podID="dd6f8ebd-b1b9-4142-8012-4310c45a041f" containerID="7c094296cbbddf544b6fd7851462d7a082546e3a64eec89121444d8f1dcbcccf" exitCode=137 Oct 01 16:20:08 crc kubenswrapper[4726]: I1001 16:20:08.926142 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dd6f8ebd-b1b9-4142-8012-4310c45a041f","Type":"ContainerDied","Data":"7c094296cbbddf544b6fd7851462d7a082546e3a64eec89121444d8f1dcbcccf"} Oct 01 16:20:08 crc kubenswrapper[4726]: I1001 16:20:08.926160 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:08 crc kubenswrapper[4726]: I1001 16:20:08.926173 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dd6f8ebd-b1b9-4142-8012-4310c45a041f","Type":"ContainerDied","Data":"f1f258a750a7d790beb723b9fb3e124727bf7f91340a474e6c8de7b63034bfba"} Oct 01 16:20:08 crc kubenswrapper[4726]: I1001 16:20:08.926192 4726 scope.go:117] "RemoveContainer" containerID="7c094296cbbddf544b6fd7851462d7a082546e3a64eec89121444d8f1dcbcccf" Oct 01 16:20:08 crc kubenswrapper[4726]: I1001 16:20:08.952436 4726 scope.go:117] "RemoveContainer" containerID="7c094296cbbddf544b6fd7851462d7a082546e3a64eec89121444d8f1dcbcccf" Oct 01 16:20:08 crc kubenswrapper[4726]: E1001 16:20:08.952993 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c094296cbbddf544b6fd7851462d7a082546e3a64eec89121444d8f1dcbcccf\": container with ID starting with 7c094296cbbddf544b6fd7851462d7a082546e3a64eec89121444d8f1dcbcccf not found: ID does not exist" containerID="7c094296cbbddf544b6fd7851462d7a082546e3a64eec89121444d8f1dcbcccf" Oct 01 16:20:08 crc kubenswrapper[4726]: I1001 16:20:08.953038 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c094296cbbddf544b6fd7851462d7a082546e3a64eec89121444d8f1dcbcccf"} err="failed to get container status \"7c094296cbbddf544b6fd7851462d7a082546e3a64eec89121444d8f1dcbcccf\": rpc error: code = NotFound desc = could not find container \"7c094296cbbddf544b6fd7851462d7a082546e3a64eec89121444d8f1dcbcccf\": container with ID starting with 7c094296cbbddf544b6fd7851462d7a082546e3a64eec89121444d8f1dcbcccf not found: ID does not exist" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.000105 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6f8ebd-b1b9-4142-8012-4310c45a041f-config-data\") pod \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\" (UID: \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\") " Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.000323 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6f8ebd-b1b9-4142-8012-4310c45a041f-combined-ca-bundle\") pod \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\" (UID: \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\") " Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.000372 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95gwn\" (UniqueName: \"kubernetes.io/projected/dd6f8ebd-b1b9-4142-8012-4310c45a041f-kube-api-access-95gwn\") pod \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\" (UID: \"dd6f8ebd-b1b9-4142-8012-4310c45a041f\") " Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.007263 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd6f8ebd-b1b9-4142-8012-4310c45a041f-kube-api-access-95gwn" (OuterVolumeSpecName: "kube-api-access-95gwn") pod "dd6f8ebd-b1b9-4142-8012-4310c45a041f" (UID: "dd6f8ebd-b1b9-4142-8012-4310c45a041f"). InnerVolumeSpecName "kube-api-access-95gwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.033371 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd6f8ebd-b1b9-4142-8012-4310c45a041f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd6f8ebd-b1b9-4142-8012-4310c45a041f" (UID: "dd6f8ebd-b1b9-4142-8012-4310c45a041f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.038373 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd6f8ebd-b1b9-4142-8012-4310c45a041f-config-data" (OuterVolumeSpecName: "config-data") pod "dd6f8ebd-b1b9-4142-8012-4310c45a041f" (UID: "dd6f8ebd-b1b9-4142-8012-4310c45a041f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.102868 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6f8ebd-b1b9-4142-8012-4310c45a041f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.103178 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6f8ebd-b1b9-4142-8012-4310c45a041f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.103193 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95gwn\" (UniqueName: \"kubernetes.io/projected/dd6f8ebd-b1b9-4142-8012-4310c45a041f-kube-api-access-95gwn\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.270530 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.281866 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.304737 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:20:09 crc kubenswrapper[4726]: E1001 16:20:09.305265 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd6f8ebd-b1b9-4142-8012-4310c45a041f" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.305290 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd6f8ebd-b1b9-4142-8012-4310c45a041f" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.305505 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd6f8ebd-b1b9-4142-8012-4310c45a041f" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.306226 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.309616 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.309682 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.310167 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.330027 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.388574 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.389515 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.390356 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.392239 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.407952 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.408008 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsqjt\" (UniqueName: \"kubernetes.io/projected/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-kube-api-access-hsqjt\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.408083 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.408171 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.408281 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.509866 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.510265 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.510398 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsqjt\" (UniqueName: \"kubernetes.io/projected/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-kube-api-access-hsqjt\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.510520 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.510733 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.514734 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.515200 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.515469 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.519287 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.534372 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsqjt\" (UniqueName: \"kubernetes.io/projected/0c2bafde-fa7b-4e17-a5da-e2f2caa63012-kube-api-access-hsqjt\") pod \"nova-cell1-novncproxy-0\" (UID: \"0c2bafde-fa7b-4e17-a5da-e2f2caa63012\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.650323 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.825256 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd6f8ebd-b1b9-4142-8012-4310c45a041f" path="/var/lib/kubelet/pods/dd6f8ebd-b1b9-4142-8012-4310c45a041f/volumes" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.936003 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 16:20:09 crc kubenswrapper[4726]: I1001 16:20:09.941316 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.140820 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-jwdxg"] Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.143187 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.151183 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-jwdxg"] Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.161504 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.224211 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.224251 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.224295 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-config\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.224316 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.224410 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tl6d\" (UniqueName: \"kubernetes.io/projected/70a049b2-bfa9-4ccf-bddb-ece60978a6be-kube-api-access-5tl6d\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.224451 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.326006 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.326069 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.326108 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-config\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.326132 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.326165 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tl6d\" (UniqueName: \"kubernetes.io/projected/70a049b2-bfa9-4ccf-bddb-ece60978a6be-kube-api-access-5tl6d\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.326188 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.327111 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.327364 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-config\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.327481 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.327504 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.327729 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.345113 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tl6d\" (UniqueName: \"kubernetes.io/projected/70a049b2-bfa9-4ccf-bddb-ece60978a6be-kube-api-access-5tl6d\") pod \"dnsmasq-dns-cd5cbd7b9-jwdxg\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.508830 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.947501 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0c2bafde-fa7b-4e17-a5da-e2f2caa63012","Type":"ContainerStarted","Data":"fbc0fbc84a707cb544255527368242f691c0beef113f2f5a4cbc21bb555d7104"} Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.947802 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0c2bafde-fa7b-4e17-a5da-e2f2caa63012","Type":"ContainerStarted","Data":"85c071999ed7a6de716c43912d636edd613f135a56e3d4ea3df92cd48916acd6"} Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.973352 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.973334979 podStartE2EDuration="1.973334979s" podCreationTimestamp="2025-10-01 16:20:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:20:10.966849056 +0000 UTC m=+1263.868401643" watchObservedRunningTime="2025-10-01 16:20:10.973334979 +0000 UTC m=+1263.874887556" Oct 01 16:20:10 crc kubenswrapper[4726]: I1001 16:20:10.994482 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-jwdxg"] Oct 01 16:20:11 crc kubenswrapper[4726]: I1001 16:20:11.959438 4726 generic.go:334] "Generic (PLEG): container finished" podID="70a049b2-bfa9-4ccf-bddb-ece60978a6be" containerID="e6dfbc8d83b3bb5bb8192b054e6ab08f74173fb11f367fdd07b7ca9d0e002008" exitCode=0 Oct 01 16:20:11 crc kubenswrapper[4726]: I1001 16:20:11.959874 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" event={"ID":"70a049b2-bfa9-4ccf-bddb-ece60978a6be","Type":"ContainerDied","Data":"e6dfbc8d83b3bb5bb8192b054e6ab08f74173fb11f367fdd07b7ca9d0e002008"} Oct 01 16:20:11 crc kubenswrapper[4726]: I1001 16:20:11.959940 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" event={"ID":"70a049b2-bfa9-4ccf-bddb-ece60978a6be","Type":"ContainerStarted","Data":"94fd35abc08c223b29a56c6d04645d91598aa032d8bc4f208b31b7ab227a6ef2"} Oct 01 16:20:12 crc kubenswrapper[4726]: I1001 16:20:12.245918 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:20:12 crc kubenswrapper[4726]: I1001 16:20:12.246511 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="ceilometer-central-agent" containerID="cri-o://b9abcb235c3f3a08d5e9ad4619b5cba68560ecc4bb9c612636baba38cee061be" gracePeriod=30 Oct 01 16:20:12 crc kubenswrapper[4726]: I1001 16:20:12.246576 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="proxy-httpd" containerID="cri-o://ed56207bdcf0dc3ac3e80204eacc9ccaadc4494bafb39c317cd690d9e298a0a0" gracePeriod=30 Oct 01 16:20:12 crc kubenswrapper[4726]: I1001 16:20:12.246601 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="sg-core" containerID="cri-o://602e5266a49092b8ddb435ec2c1bc3d6c770ccbd5be1058f3371c6c20f34fd93" gracePeriod=30 Oct 01 16:20:12 crc kubenswrapper[4726]: I1001 16:20:12.246613 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="ceilometer-notification-agent" containerID="cri-o://a0f250ba81efd279b348c1a80272ccb11c648ac1542741ec2633d1bb5eed421b" gracePeriod=30 Oct 01 16:20:12 crc kubenswrapper[4726]: I1001 16:20:12.538429 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:20:12 crc kubenswrapper[4726]: I1001 16:20:12.992908 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" event={"ID":"70a049b2-bfa9-4ccf-bddb-ece60978a6be","Type":"ContainerStarted","Data":"f13efaa67ae6b7e10b323aa5a532f9cff48c69cf9d95c5cfded2a150b5cc4531"} Oct 01 16:20:12 crc kubenswrapper[4726]: I1001 16:20:12.992974 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:13 crc kubenswrapper[4726]: I1001 16:20:13.005507 4726 generic.go:334] "Generic (PLEG): container finished" podID="fdf946f7-e990-4849-a998-3d151a619bf2" containerID="ed56207bdcf0dc3ac3e80204eacc9ccaadc4494bafb39c317cd690d9e298a0a0" exitCode=0 Oct 01 16:20:13 crc kubenswrapper[4726]: I1001 16:20:13.005540 4726 generic.go:334] "Generic (PLEG): container finished" podID="fdf946f7-e990-4849-a998-3d151a619bf2" containerID="602e5266a49092b8ddb435ec2c1bc3d6c770ccbd5be1058f3371c6c20f34fd93" exitCode=2 Oct 01 16:20:13 crc kubenswrapper[4726]: I1001 16:20:13.005548 4726 generic.go:334] "Generic (PLEG): container finished" podID="fdf946f7-e990-4849-a998-3d151a619bf2" containerID="b9abcb235c3f3a08d5e9ad4619b5cba68560ecc4bb9c612636baba38cee061be" exitCode=0 Oct 01 16:20:13 crc kubenswrapper[4726]: I1001 16:20:13.005925 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" containerName="nova-api-log" containerID="cri-o://fcb12de2fb93ae3828d3e48cc8c8cde46249cf78784ab46772f8d1650f7a86bb" gracePeriod=30 Oct 01 16:20:13 crc kubenswrapper[4726]: I1001 16:20:13.006762 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fdf946f7-e990-4849-a998-3d151a619bf2","Type":"ContainerDied","Data":"ed56207bdcf0dc3ac3e80204eacc9ccaadc4494bafb39c317cd690d9e298a0a0"} Oct 01 16:20:13 crc kubenswrapper[4726]: I1001 16:20:13.006832 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fdf946f7-e990-4849-a998-3d151a619bf2","Type":"ContainerDied","Data":"602e5266a49092b8ddb435ec2c1bc3d6c770ccbd5be1058f3371c6c20f34fd93"} Oct 01 16:20:13 crc kubenswrapper[4726]: I1001 16:20:13.006849 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fdf946f7-e990-4849-a998-3d151a619bf2","Type":"ContainerDied","Data":"b9abcb235c3f3a08d5e9ad4619b5cba68560ecc4bb9c612636baba38cee061be"} Oct 01 16:20:13 crc kubenswrapper[4726]: I1001 16:20:13.006966 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" containerName="nova-api-api" containerID="cri-o://9698c8bd47f1703eb9e267cfc613890b07bf7545abe58ab8c41065f3e6e9c86e" gracePeriod=30 Oct 01 16:20:13 crc kubenswrapper[4726]: I1001 16:20:13.013495 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" podStartSLOduration=3.013478026 podStartE2EDuration="3.013478026s" podCreationTimestamp="2025-10-01 16:20:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:20:13.010818131 +0000 UTC m=+1265.912370708" watchObservedRunningTime="2025-10-01 16:20:13.013478026 +0000 UTC m=+1265.915030603" Oct 01 16:20:14 crc kubenswrapper[4726]: I1001 16:20:14.014715 4726 generic.go:334] "Generic (PLEG): container finished" podID="78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" containerID="fcb12de2fb93ae3828d3e48cc8c8cde46249cf78784ab46772f8d1650f7a86bb" exitCode=143 Oct 01 16:20:14 crc kubenswrapper[4726]: I1001 16:20:14.014797 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76","Type":"ContainerDied","Data":"fcb12de2fb93ae3828d3e48cc8c8cde46249cf78784ab46772f8d1650f7a86bb"} Oct 01 16:20:14 crc kubenswrapper[4726]: I1001 16:20:14.651263 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:16 crc kubenswrapper[4726]: I1001 16:20:16.624165 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:20:16 crc kubenswrapper[4726]: I1001 16:20:16.778889 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-combined-ca-bundle\") pod \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " Oct 01 16:20:16 crc kubenswrapper[4726]: I1001 16:20:16.779526 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phsnk\" (UniqueName: \"kubernetes.io/projected/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-kube-api-access-phsnk\") pod \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " Oct 01 16:20:16 crc kubenswrapper[4726]: I1001 16:20:16.779587 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-logs\") pod \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " Oct 01 16:20:16 crc kubenswrapper[4726]: I1001 16:20:16.779640 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-config-data\") pod \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\" (UID: \"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76\") " Oct 01 16:20:16 crc kubenswrapper[4726]: I1001 16:20:16.780101 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-logs" (OuterVolumeSpecName: "logs") pod "78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" (UID: "78f1d4ce-6d8b-4e71-8ceb-311efebc9d76"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:20:16 crc kubenswrapper[4726]: I1001 16:20:16.780264 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:16 crc kubenswrapper[4726]: I1001 16:20:16.787457 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-kube-api-access-phsnk" (OuterVolumeSpecName: "kube-api-access-phsnk") pod "78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" (UID: "78f1d4ce-6d8b-4e71-8ceb-311efebc9d76"). InnerVolumeSpecName "kube-api-access-phsnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:20:16 crc kubenswrapper[4726]: I1001 16:20:16.818963 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" (UID: "78f1d4ce-6d8b-4e71-8ceb-311efebc9d76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:16 crc kubenswrapper[4726]: I1001 16:20:16.823675 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-config-data" (OuterVolumeSpecName: "config-data") pod "78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" (UID: "78f1d4ce-6d8b-4e71-8ceb-311efebc9d76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:16 crc kubenswrapper[4726]: I1001 16:20:16.882584 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:16 crc kubenswrapper[4726]: I1001 16:20:16.882623 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phsnk\" (UniqueName: \"kubernetes.io/projected/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-kube-api-access-phsnk\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:16 crc kubenswrapper[4726]: I1001 16:20:16.882679 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:16 crc kubenswrapper[4726]: I1001 16:20:16.970482 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.041070 4726 generic.go:334] "Generic (PLEG): container finished" podID="78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" containerID="9698c8bd47f1703eb9e267cfc613890b07bf7545abe58ab8c41065f3e6e9c86e" exitCode=0 Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.041134 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76","Type":"ContainerDied","Data":"9698c8bd47f1703eb9e267cfc613890b07bf7545abe58ab8c41065f3e6e9c86e"} Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.041158 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78f1d4ce-6d8b-4e71-8ceb-311efebc9d76","Type":"ContainerDied","Data":"0e33d9ef7fa6bf6e792497953f4fb080826deffb77b0ee3ec3d60ecec6272bd7"} Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.041174 4726 scope.go:117] "RemoveContainer" containerID="9698c8bd47f1703eb9e267cfc613890b07bf7545abe58ab8c41065f3e6e9c86e" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.041192 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.046387 4726 generic.go:334] "Generic (PLEG): container finished" podID="fdf946f7-e990-4849-a998-3d151a619bf2" containerID="a0f250ba81efd279b348c1a80272ccb11c648ac1542741ec2633d1bb5eed421b" exitCode=0 Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.046432 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fdf946f7-e990-4849-a998-3d151a619bf2","Type":"ContainerDied","Data":"a0f250ba81efd279b348c1a80272ccb11c648ac1542741ec2633d1bb5eed421b"} Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.046463 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fdf946f7-e990-4849-a998-3d151a619bf2","Type":"ContainerDied","Data":"fb4ec01d0824f3dd1bd3fb39cc5f104318934745df1c8604a1f0cc976ecf2485"} Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.046543 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.075531 4726 scope.go:117] "RemoveContainer" containerID="fcb12de2fb93ae3828d3e48cc8c8cde46249cf78784ab46772f8d1650f7a86bb" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.079714 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.087208 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-scripts\") pod \"fdf946f7-e990-4849-a998-3d151a619bf2\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.087296 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fdf946f7-e990-4849-a998-3d151a619bf2-run-httpd\") pod \"fdf946f7-e990-4849-a998-3d151a619bf2\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.087404 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-sg-core-conf-yaml\") pod \"fdf946f7-e990-4849-a998-3d151a619bf2\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.087444 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fdf946f7-e990-4849-a998-3d151a619bf2-log-httpd\") pod \"fdf946f7-e990-4849-a998-3d151a619bf2\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.087462 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-ceilometer-tls-certs\") pod \"fdf946f7-e990-4849-a998-3d151a619bf2\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.087481 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-config-data\") pod \"fdf946f7-e990-4849-a998-3d151a619bf2\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.087570 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7sfp9\" (UniqueName: \"kubernetes.io/projected/fdf946f7-e990-4849-a998-3d151a619bf2-kube-api-access-7sfp9\") pod \"fdf946f7-e990-4849-a998-3d151a619bf2\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.087588 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-combined-ca-bundle\") pod \"fdf946f7-e990-4849-a998-3d151a619bf2\" (UID: \"fdf946f7-e990-4849-a998-3d151a619bf2\") " Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.088033 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdf946f7-e990-4849-a998-3d151a619bf2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fdf946f7-e990-4849-a998-3d151a619bf2" (UID: "fdf946f7-e990-4849-a998-3d151a619bf2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.088299 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdf946f7-e990-4849-a998-3d151a619bf2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fdf946f7-e990-4849-a998-3d151a619bf2" (UID: "fdf946f7-e990-4849-a998-3d151a619bf2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.093106 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdf946f7-e990-4849-a998-3d151a619bf2-kube-api-access-7sfp9" (OuterVolumeSpecName: "kube-api-access-7sfp9") pod "fdf946f7-e990-4849-a998-3d151a619bf2" (UID: "fdf946f7-e990-4849-a998-3d151a619bf2"). InnerVolumeSpecName "kube-api-access-7sfp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.101430 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-scripts" (OuterVolumeSpecName: "scripts") pod "fdf946f7-e990-4849-a998-3d151a619bf2" (UID: "fdf946f7-e990-4849-a998-3d151a619bf2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.102754 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.112148 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.112251 4726 scope.go:117] "RemoveContainer" containerID="9698c8bd47f1703eb9e267cfc613890b07bf7545abe58ab8c41065f3e6e9c86e" Oct 01 16:20:17 crc kubenswrapper[4726]: E1001 16:20:17.112553 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9698c8bd47f1703eb9e267cfc613890b07bf7545abe58ab8c41065f3e6e9c86e\": container with ID starting with 9698c8bd47f1703eb9e267cfc613890b07bf7545abe58ab8c41065f3e6e9c86e not found: ID does not exist" containerID="9698c8bd47f1703eb9e267cfc613890b07bf7545abe58ab8c41065f3e6e9c86e" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.112587 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9698c8bd47f1703eb9e267cfc613890b07bf7545abe58ab8c41065f3e6e9c86e"} err="failed to get container status \"9698c8bd47f1703eb9e267cfc613890b07bf7545abe58ab8c41065f3e6e9c86e\": rpc error: code = NotFound desc = could not find container \"9698c8bd47f1703eb9e267cfc613890b07bf7545abe58ab8c41065f3e6e9c86e\": container with ID starting with 9698c8bd47f1703eb9e267cfc613890b07bf7545abe58ab8c41065f3e6e9c86e not found: ID does not exist" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.112614 4726 scope.go:117] "RemoveContainer" containerID="fcb12de2fb93ae3828d3e48cc8c8cde46249cf78784ab46772f8d1650f7a86bb" Oct 01 16:20:17 crc kubenswrapper[4726]: E1001 16:20:17.112627 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" containerName="nova-api-log" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.112644 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" containerName="nova-api-log" Oct 01 16:20:17 crc kubenswrapper[4726]: E1001 16:20:17.112656 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" containerName="nova-api-api" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.112665 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" containerName="nova-api-api" Oct 01 16:20:17 crc kubenswrapper[4726]: E1001 16:20:17.112707 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="sg-core" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.112716 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="sg-core" Oct 01 16:20:17 crc kubenswrapper[4726]: E1001 16:20:17.112752 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="proxy-httpd" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.112760 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="proxy-httpd" Oct 01 16:20:17 crc kubenswrapper[4726]: E1001 16:20:17.112789 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="ceilometer-central-agent" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.112798 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="ceilometer-central-agent" Oct 01 16:20:17 crc kubenswrapper[4726]: E1001 16:20:17.112795 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcb12de2fb93ae3828d3e48cc8c8cde46249cf78784ab46772f8d1650f7a86bb\": container with ID starting with fcb12de2fb93ae3828d3e48cc8c8cde46249cf78784ab46772f8d1650f7a86bb not found: ID does not exist" containerID="fcb12de2fb93ae3828d3e48cc8c8cde46249cf78784ab46772f8d1650f7a86bb" Oct 01 16:20:17 crc kubenswrapper[4726]: E1001 16:20:17.112812 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="ceilometer-notification-agent" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.112820 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="ceilometer-notification-agent" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.112816 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcb12de2fb93ae3828d3e48cc8c8cde46249cf78784ab46772f8d1650f7a86bb"} err="failed to get container status \"fcb12de2fb93ae3828d3e48cc8c8cde46249cf78784ab46772f8d1650f7a86bb\": rpc error: code = NotFound desc = could not find container \"fcb12de2fb93ae3828d3e48cc8c8cde46249cf78784ab46772f8d1650f7a86bb\": container with ID starting with fcb12de2fb93ae3828d3e48cc8c8cde46249cf78784ab46772f8d1650f7a86bb not found: ID does not exist" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.112925 4726 scope.go:117] "RemoveContainer" containerID="ed56207bdcf0dc3ac3e80204eacc9ccaadc4494bafb39c317cd690d9e298a0a0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.113035 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="sg-core" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.113085 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="ceilometer-central-agent" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.113105 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" containerName="nova-api-log" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.113115 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="proxy-httpd" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.113134 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" containerName="nova-api-api" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.113142 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" containerName="ceilometer-notification-agent" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.114504 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.117208 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.118312 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.118347 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.141427 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.147096 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fdf946f7-e990-4849-a998-3d151a619bf2" (UID: "fdf946f7-e990-4849-a998-3d151a619bf2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.149277 4726 scope.go:117] "RemoveContainer" containerID="602e5266a49092b8ddb435ec2c1bc3d6c770ccbd5be1058f3371c6c20f34fd93" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.168544 4726 scope.go:117] "RemoveContainer" containerID="a0f250ba81efd279b348c1a80272ccb11c648ac1542741ec2633d1bb5eed421b" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.169209 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "fdf946f7-e990-4849-a998-3d151a619bf2" (UID: "fdf946f7-e990-4849-a998-3d151a619bf2"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.191065 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.191093 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fdf946f7-e990-4849-a998-3d151a619bf2-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.191103 4726 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.191120 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7sfp9\" (UniqueName: \"kubernetes.io/projected/fdf946f7-e990-4849-a998-3d151a619bf2-kube-api-access-7sfp9\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.191131 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.191141 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fdf946f7-e990-4849-a998-3d151a619bf2-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.200245 4726 scope.go:117] "RemoveContainer" containerID="b9abcb235c3f3a08d5e9ad4619b5cba68560ecc4bb9c612636baba38cee061be" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.217273 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-config-data" (OuterVolumeSpecName: "config-data") pod "fdf946f7-e990-4849-a998-3d151a619bf2" (UID: "fdf946f7-e990-4849-a998-3d151a619bf2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.225618 4726 scope.go:117] "RemoveContainer" containerID="ed56207bdcf0dc3ac3e80204eacc9ccaadc4494bafb39c317cd690d9e298a0a0" Oct 01 16:20:17 crc kubenswrapper[4726]: E1001 16:20:17.225995 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed56207bdcf0dc3ac3e80204eacc9ccaadc4494bafb39c317cd690d9e298a0a0\": container with ID starting with ed56207bdcf0dc3ac3e80204eacc9ccaadc4494bafb39c317cd690d9e298a0a0 not found: ID does not exist" containerID="ed56207bdcf0dc3ac3e80204eacc9ccaadc4494bafb39c317cd690d9e298a0a0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.226039 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed56207bdcf0dc3ac3e80204eacc9ccaadc4494bafb39c317cd690d9e298a0a0"} err="failed to get container status \"ed56207bdcf0dc3ac3e80204eacc9ccaadc4494bafb39c317cd690d9e298a0a0\": rpc error: code = NotFound desc = could not find container \"ed56207bdcf0dc3ac3e80204eacc9ccaadc4494bafb39c317cd690d9e298a0a0\": container with ID starting with ed56207bdcf0dc3ac3e80204eacc9ccaadc4494bafb39c317cd690d9e298a0a0 not found: ID does not exist" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.226087 4726 scope.go:117] "RemoveContainer" containerID="602e5266a49092b8ddb435ec2c1bc3d6c770ccbd5be1058f3371c6c20f34fd93" Oct 01 16:20:17 crc kubenswrapper[4726]: E1001 16:20:17.228899 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"602e5266a49092b8ddb435ec2c1bc3d6c770ccbd5be1058f3371c6c20f34fd93\": container with ID starting with 602e5266a49092b8ddb435ec2c1bc3d6c770ccbd5be1058f3371c6c20f34fd93 not found: ID does not exist" containerID="602e5266a49092b8ddb435ec2c1bc3d6c770ccbd5be1058f3371c6c20f34fd93" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.228937 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"602e5266a49092b8ddb435ec2c1bc3d6c770ccbd5be1058f3371c6c20f34fd93"} err="failed to get container status \"602e5266a49092b8ddb435ec2c1bc3d6c770ccbd5be1058f3371c6c20f34fd93\": rpc error: code = NotFound desc = could not find container \"602e5266a49092b8ddb435ec2c1bc3d6c770ccbd5be1058f3371c6c20f34fd93\": container with ID starting with 602e5266a49092b8ddb435ec2c1bc3d6c770ccbd5be1058f3371c6c20f34fd93 not found: ID does not exist" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.228961 4726 scope.go:117] "RemoveContainer" containerID="a0f250ba81efd279b348c1a80272ccb11c648ac1542741ec2633d1bb5eed421b" Oct 01 16:20:17 crc kubenswrapper[4726]: E1001 16:20:17.229652 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0f250ba81efd279b348c1a80272ccb11c648ac1542741ec2633d1bb5eed421b\": container with ID starting with a0f250ba81efd279b348c1a80272ccb11c648ac1542741ec2633d1bb5eed421b not found: ID does not exist" containerID="a0f250ba81efd279b348c1a80272ccb11c648ac1542741ec2633d1bb5eed421b" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.229706 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0f250ba81efd279b348c1a80272ccb11c648ac1542741ec2633d1bb5eed421b"} err="failed to get container status \"a0f250ba81efd279b348c1a80272ccb11c648ac1542741ec2633d1bb5eed421b\": rpc error: code = NotFound desc = could not find container \"a0f250ba81efd279b348c1a80272ccb11c648ac1542741ec2633d1bb5eed421b\": container with ID starting with a0f250ba81efd279b348c1a80272ccb11c648ac1542741ec2633d1bb5eed421b not found: ID does not exist" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.229735 4726 scope.go:117] "RemoveContainer" containerID="b9abcb235c3f3a08d5e9ad4619b5cba68560ecc4bb9c612636baba38cee061be" Oct 01 16:20:17 crc kubenswrapper[4726]: E1001 16:20:17.230253 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9abcb235c3f3a08d5e9ad4619b5cba68560ecc4bb9c612636baba38cee061be\": container with ID starting with b9abcb235c3f3a08d5e9ad4619b5cba68560ecc4bb9c612636baba38cee061be not found: ID does not exist" containerID="b9abcb235c3f3a08d5e9ad4619b5cba68560ecc4bb9c612636baba38cee061be" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.230391 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9abcb235c3f3a08d5e9ad4619b5cba68560ecc4bb9c612636baba38cee061be"} err="failed to get container status \"b9abcb235c3f3a08d5e9ad4619b5cba68560ecc4bb9c612636baba38cee061be\": rpc error: code = NotFound desc = could not find container \"b9abcb235c3f3a08d5e9ad4619b5cba68560ecc4bb9c612636baba38cee061be\": container with ID starting with b9abcb235c3f3a08d5e9ad4619b5cba68560ecc4bb9c612636baba38cee061be not found: ID does not exist" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.230322 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fdf946f7-e990-4849-a998-3d151a619bf2" (UID: "fdf946f7-e990-4849-a998-3d151a619bf2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.292809 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8vcx\" (UniqueName: \"kubernetes.io/projected/8000eb48-af65-45e9-abe6-3263671bc050-kube-api-access-j8vcx\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.292913 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.292940 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-config-data\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.292988 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-public-tls-certs\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.293026 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8000eb48-af65-45e9-abe6-3263671bc050-logs\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.293103 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.293249 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.293261 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdf946f7-e990-4849-a998-3d151a619bf2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.377294 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.386477 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.395374 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8vcx\" (UniqueName: \"kubernetes.io/projected/8000eb48-af65-45e9-abe6-3263671bc050-kube-api-access-j8vcx\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.395423 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.395454 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-config-data\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.395468 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-public-tls-certs\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.395512 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8000eb48-af65-45e9-abe6-3263671bc050-logs\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.395560 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.399334 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.401170 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.404034 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.407323 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.407855 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-public-tls-certs\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.408265 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-config-data\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.408523 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.408562 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8000eb48-af65-45e9-abe6-3263671bc050-logs\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.415209 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.415418 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.417360 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.431030 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8vcx\" (UniqueName: \"kubernetes.io/projected/8000eb48-af65-45e9-abe6-3263671bc050-kube-api-access-j8vcx\") pod \"nova-api-0\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.437935 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.497795 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.498106 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-575pj\" (UniqueName: \"kubernetes.io/projected/0a303d6a-38ee-4641-b10b-d2627121df03-kube-api-access-575pj\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.498247 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.498543 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a303d6a-38ee-4641-b10b-d2627121df03-log-httpd\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.498650 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-config-data\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.498695 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.498725 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a303d6a-38ee-4641-b10b-d2627121df03-run-httpd\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.498777 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-scripts\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.600586 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.600691 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a303d6a-38ee-4641-b10b-d2627121df03-log-httpd\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.600735 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-config-data\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.600766 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.600789 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a303d6a-38ee-4641-b10b-d2627121df03-run-httpd\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.600818 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-scripts\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.600864 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.600897 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-575pj\" (UniqueName: \"kubernetes.io/projected/0a303d6a-38ee-4641-b10b-d2627121df03-kube-api-access-575pj\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.601692 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a303d6a-38ee-4641-b10b-d2627121df03-log-httpd\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.601931 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a303d6a-38ee-4641-b10b-d2627121df03-run-httpd\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.604748 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.605144 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.605746 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-scripts\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.606193 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.607489 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a303d6a-38ee-4641-b10b-d2627121df03-config-data\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.616524 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-575pj\" (UniqueName: \"kubernetes.io/projected/0a303d6a-38ee-4641-b10b-d2627121df03-kube-api-access-575pj\") pod \"ceilometer-0\" (UID: \"0a303d6a-38ee-4641-b10b-d2627121df03\") " pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.820311 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78f1d4ce-6d8b-4e71-8ceb-311efebc9d76" path="/var/lib/kubelet/pods/78f1d4ce-6d8b-4e71-8ceb-311efebc9d76/volumes" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.821087 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdf946f7-e990-4849-a998-3d151a619bf2" path="/var/lib/kubelet/pods/fdf946f7-e990-4849-a998-3d151a619bf2/volumes" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.876336 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:20:17 crc kubenswrapper[4726]: I1001 16:20:17.920348 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:20:18 crc kubenswrapper[4726]: I1001 16:20:18.058018 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8000eb48-af65-45e9-abe6-3263671bc050","Type":"ContainerStarted","Data":"22254b36c44abe58270a2a399e0c47dc416cb63506be13393ac11f01b692cf9c"} Oct 01 16:20:19 crc kubenswrapper[4726]: I1001 16:20:19.074785 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8000eb48-af65-45e9-abe6-3263671bc050","Type":"ContainerStarted","Data":"030a821c9f34415d3a9c8a9e410bfc407619d81d38dcc9cb1e7a5422ece5b681"} Oct 01 16:20:19 crc kubenswrapper[4726]: I1001 16:20:19.076191 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8000eb48-af65-45e9-abe6-3263671bc050","Type":"ContainerStarted","Data":"e586999f7364dfff65f561c4b34b9b526d9f937fa6053ef6a56d0d73904815a1"} Oct 01 16:20:19 crc kubenswrapper[4726]: I1001 16:20:19.102975 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.102954747 podStartE2EDuration="2.102954747s" podCreationTimestamp="2025-10-01 16:20:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:20:19.093646465 +0000 UTC m=+1271.995199062" watchObservedRunningTime="2025-10-01 16:20:19.102954747 +0000 UTC m=+1272.004507324" Oct 01 16:20:19 crc kubenswrapper[4726]: I1001 16:20:19.128577 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:20:19 crc kubenswrapper[4726]: W1001 16:20:19.131872 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a303d6a_38ee_4641_b10b_d2627121df03.slice/crio-7d064a0fa8d55c047aae3f40bbdc7feb6ee87e664b5f8d94d46395ed46ad4a01 WatchSource:0}: Error finding container 7d064a0fa8d55c047aae3f40bbdc7feb6ee87e664b5f8d94d46395ed46ad4a01: Status 404 returned error can't find the container with id 7d064a0fa8d55c047aae3f40bbdc7feb6ee87e664b5f8d94d46395ed46ad4a01 Oct 01 16:20:19 crc kubenswrapper[4726]: I1001 16:20:19.134797 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:20:19 crc kubenswrapper[4726]: I1001 16:20:19.651020 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:19 crc kubenswrapper[4726]: I1001 16:20:19.670655 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.085554 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a303d6a-38ee-4641-b10b-d2627121df03","Type":"ContainerStarted","Data":"46e423945df88e519d0643d674159488e598020e7f05d8537545afaaf9e7ab85"} Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.085835 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a303d6a-38ee-4641-b10b-d2627121df03","Type":"ContainerStarted","Data":"7d064a0fa8d55c047aae3f40bbdc7feb6ee87e664b5f8d94d46395ed46ad4a01"} Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.101425 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.267188 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-vj74b"] Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.269355 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.271549 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.272756 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.277780 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-vj74b"] Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.462900 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-scripts\") pod \"nova-cell1-cell-mapping-vj74b\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.463230 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78zn8\" (UniqueName: \"kubernetes.io/projected/5033e1a8-ab77-4894-ac5b-95ca39e731fa-kube-api-access-78zn8\") pod \"nova-cell1-cell-mapping-vj74b\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.463260 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-config-data\") pod \"nova-cell1-cell-mapping-vj74b\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.463297 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vj74b\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.515554 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.564637 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-config-data\") pod \"nova-cell1-cell-mapping-vj74b\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.564699 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vj74b\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.564784 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-scripts\") pod \"nova-cell1-cell-mapping-vj74b\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.564879 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78zn8\" (UniqueName: \"kubernetes.io/projected/5033e1a8-ab77-4894-ac5b-95ca39e731fa-kube-api-access-78zn8\") pod \"nova-cell1-cell-mapping-vj74b\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.571298 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-config-data\") pod \"nova-cell1-cell-mapping-vj74b\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.571674 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-scripts\") pod \"nova-cell1-cell-mapping-vj74b\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.573009 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vj74b\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.595078 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78zn8\" (UniqueName: \"kubernetes.io/projected/5033e1a8-ab77-4894-ac5b-95ca39e731fa-kube-api-access-78zn8\") pod \"nova-cell1-cell-mapping-vj74b\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.612502 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-md5d8"] Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.612776 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-md5d8" podUID="60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" containerName="dnsmasq-dns" containerID="cri-o://1b0d9345e2119c42c63000a1a380db0f692b70f9faee7873767ae62a1654c858" gracePeriod=10 Oct 01 16:20:20 crc kubenswrapper[4726]: I1001 16:20:20.894469 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.068362 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.113701 4726 generic.go:334] "Generic (PLEG): container finished" podID="60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" containerID="1b0d9345e2119c42c63000a1a380db0f692b70f9faee7873767ae62a1654c858" exitCode=0 Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.113767 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-md5d8" event={"ID":"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6","Type":"ContainerDied","Data":"1b0d9345e2119c42c63000a1a380db0f692b70f9faee7873767ae62a1654c858"} Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.113795 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-md5d8" event={"ID":"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6","Type":"ContainerDied","Data":"086eec88302b30529260a6b184e4c77826e6178cb3069fc4a7ddf8158a4b7ea8"} Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.113815 4726 scope.go:117] "RemoveContainer" containerID="1b0d9345e2119c42c63000a1a380db0f692b70f9faee7873767ae62a1654c858" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.113939 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-md5d8" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.125384 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a303d6a-38ee-4641-b10b-d2627121df03","Type":"ContainerStarted","Data":"38fac63005220d20513254a91ef6ad06d2c3789f73c8649c6fc742c21606f499"} Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.184710 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmv5x\" (UniqueName: \"kubernetes.io/projected/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-kube-api-access-nmv5x\") pod \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.184773 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-config\") pod \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.184971 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-ovsdbserver-sb\") pod \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.185000 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-dns-svc\") pod \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.185027 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-dns-swift-storage-0\") pod \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.185094 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-ovsdbserver-nb\") pod \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\" (UID: \"60e46f07-6ffb-46ca-8d0d-a5749dfd42f6\") " Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.192956 4726 scope.go:117] "RemoveContainer" containerID="38c5a02a0e7abe05039101d8539cf994fff94889411917ed1bc64191d7442572" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.193503 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-kube-api-access-nmv5x" (OuterVolumeSpecName: "kube-api-access-nmv5x") pod "60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" (UID: "60e46f07-6ffb-46ca-8d0d-a5749dfd42f6"). InnerVolumeSpecName "kube-api-access-nmv5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.267584 4726 scope.go:117] "RemoveContainer" containerID="1b0d9345e2119c42c63000a1a380db0f692b70f9faee7873767ae62a1654c858" Oct 01 16:20:21 crc kubenswrapper[4726]: E1001 16:20:21.271132 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b0d9345e2119c42c63000a1a380db0f692b70f9faee7873767ae62a1654c858\": container with ID starting with 1b0d9345e2119c42c63000a1a380db0f692b70f9faee7873767ae62a1654c858 not found: ID does not exist" containerID="1b0d9345e2119c42c63000a1a380db0f692b70f9faee7873767ae62a1654c858" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.271236 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b0d9345e2119c42c63000a1a380db0f692b70f9faee7873767ae62a1654c858"} err="failed to get container status \"1b0d9345e2119c42c63000a1a380db0f692b70f9faee7873767ae62a1654c858\": rpc error: code = NotFound desc = could not find container \"1b0d9345e2119c42c63000a1a380db0f692b70f9faee7873767ae62a1654c858\": container with ID starting with 1b0d9345e2119c42c63000a1a380db0f692b70f9faee7873767ae62a1654c858 not found: ID does not exist" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.271349 4726 scope.go:117] "RemoveContainer" containerID="38c5a02a0e7abe05039101d8539cf994fff94889411917ed1bc64191d7442572" Oct 01 16:20:21 crc kubenswrapper[4726]: E1001 16:20:21.275127 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38c5a02a0e7abe05039101d8539cf994fff94889411917ed1bc64191d7442572\": container with ID starting with 38c5a02a0e7abe05039101d8539cf994fff94889411917ed1bc64191d7442572 not found: ID does not exist" containerID="38c5a02a0e7abe05039101d8539cf994fff94889411917ed1bc64191d7442572" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.275194 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38c5a02a0e7abe05039101d8539cf994fff94889411917ed1bc64191d7442572"} err="failed to get container status \"38c5a02a0e7abe05039101d8539cf994fff94889411917ed1bc64191d7442572\": rpc error: code = NotFound desc = could not find container \"38c5a02a0e7abe05039101d8539cf994fff94889411917ed1bc64191d7442572\": container with ID starting with 38c5a02a0e7abe05039101d8539cf994fff94889411917ed1bc64191d7442572 not found: ID does not exist" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.287215 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmv5x\" (UniqueName: \"kubernetes.io/projected/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-kube-api-access-nmv5x\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.301042 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-config" (OuterVolumeSpecName: "config") pod "60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" (UID: "60e46f07-6ffb-46ca-8d0d-a5749dfd42f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.301980 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" (UID: "60e46f07-6ffb-46ca-8d0d-a5749dfd42f6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.304654 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" (UID: "60e46f07-6ffb-46ca-8d0d-a5749dfd42f6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.306545 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" (UID: "60e46f07-6ffb-46ca-8d0d-a5749dfd42f6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.317520 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" (UID: "60e46f07-6ffb-46ca-8d0d-a5749dfd42f6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.388857 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.388896 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.388905 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.388914 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.388927 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.451791 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-md5d8"] Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.463343 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-md5d8"] Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.560123 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-vj74b"] Oct 01 16:20:21 crc kubenswrapper[4726]: W1001 16:20:21.562651 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5033e1a8_ab77_4894_ac5b_95ca39e731fa.slice/crio-d963a69db3db118290eb496b50bcafe982e4403fa72433d981c2e94721cdf69f WatchSource:0}: Error finding container d963a69db3db118290eb496b50bcafe982e4403fa72433d981c2e94721cdf69f: Status 404 returned error can't find the container with id d963a69db3db118290eb496b50bcafe982e4403fa72433d981c2e94721cdf69f Oct 01 16:20:21 crc kubenswrapper[4726]: I1001 16:20:21.820747 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" path="/var/lib/kubelet/pods/60e46f07-6ffb-46ca-8d0d-a5749dfd42f6/volumes" Oct 01 16:20:22 crc kubenswrapper[4726]: I1001 16:20:22.136629 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a303d6a-38ee-4641-b10b-d2627121df03","Type":"ContainerStarted","Data":"c61616a7a08abf6b79b314396e541b5b05dad9f95bd21e604b3651ce8aceaf3b"} Oct 01 16:20:22 crc kubenswrapper[4726]: I1001 16:20:22.139143 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vj74b" event={"ID":"5033e1a8-ab77-4894-ac5b-95ca39e731fa","Type":"ContainerStarted","Data":"630c37bbb33455a45b1327918f394cbce39071d19b9ee391447795f0d1909969"} Oct 01 16:20:22 crc kubenswrapper[4726]: I1001 16:20:22.139245 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vj74b" event={"ID":"5033e1a8-ab77-4894-ac5b-95ca39e731fa","Type":"ContainerStarted","Data":"d963a69db3db118290eb496b50bcafe982e4403fa72433d981c2e94721cdf69f"} Oct 01 16:20:22 crc kubenswrapper[4726]: I1001 16:20:22.165335 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-vj74b" podStartSLOduration=2.165311154 podStartE2EDuration="2.165311154s" podCreationTimestamp="2025-10-01 16:20:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:20:22.154830648 +0000 UTC m=+1275.056383235" watchObservedRunningTime="2025-10-01 16:20:22.165311154 +0000 UTC m=+1275.066863751" Oct 01 16:20:24 crc kubenswrapper[4726]: I1001 16:20:24.162698 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a303d6a-38ee-4641-b10b-d2627121df03","Type":"ContainerStarted","Data":"36d7ebd6ed3a5e45e5deba57acf560289f53d1bb59f9b7c31efcc0c67c5dcb2b"} Oct 01 16:20:24 crc kubenswrapper[4726]: I1001 16:20:24.164770 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 16:20:24 crc kubenswrapper[4726]: I1001 16:20:24.216533 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.274802323 podStartE2EDuration="7.216507292s" podCreationTimestamp="2025-10-01 16:20:17 +0000 UTC" firstStartedPulling="2025-10-01 16:20:19.134589659 +0000 UTC m=+1272.036142236" lastFinishedPulling="2025-10-01 16:20:23.076294628 +0000 UTC m=+1275.977847205" observedRunningTime="2025-10-01 16:20:24.203143656 +0000 UTC m=+1277.104696263" watchObservedRunningTime="2025-10-01 16:20:24.216507292 +0000 UTC m=+1277.118059879" Oct 01 16:20:26 crc kubenswrapper[4726]: I1001 16:20:26.001627 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-bccf8f775-md5d8" podUID="60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.189:5353: i/o timeout" Oct 01 16:20:27 crc kubenswrapper[4726]: I1001 16:20:27.210340 4726 generic.go:334] "Generic (PLEG): container finished" podID="5033e1a8-ab77-4894-ac5b-95ca39e731fa" containerID="630c37bbb33455a45b1327918f394cbce39071d19b9ee391447795f0d1909969" exitCode=0 Oct 01 16:20:27 crc kubenswrapper[4726]: I1001 16:20:27.210432 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vj74b" event={"ID":"5033e1a8-ab77-4894-ac5b-95ca39e731fa","Type":"ContainerDied","Data":"630c37bbb33455a45b1327918f394cbce39071d19b9ee391447795f0d1909969"} Oct 01 16:20:27 crc kubenswrapper[4726]: I1001 16:20:27.438800 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:20:27 crc kubenswrapper[4726]: I1001 16:20:27.438852 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.454316 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8000eb48-af65-45e9-abe6-3263671bc050" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.454338 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8000eb48-af65-45e9-abe6-3263671bc050" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.613564 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.755450 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-scripts\") pod \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.755487 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-combined-ca-bundle\") pod \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.755510 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78zn8\" (UniqueName: \"kubernetes.io/projected/5033e1a8-ab77-4894-ac5b-95ca39e731fa-kube-api-access-78zn8\") pod \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.755532 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-config-data\") pod \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\" (UID: \"5033e1a8-ab77-4894-ac5b-95ca39e731fa\") " Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.761416 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5033e1a8-ab77-4894-ac5b-95ca39e731fa-kube-api-access-78zn8" (OuterVolumeSpecName: "kube-api-access-78zn8") pod "5033e1a8-ab77-4894-ac5b-95ca39e731fa" (UID: "5033e1a8-ab77-4894-ac5b-95ca39e731fa"). InnerVolumeSpecName "kube-api-access-78zn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.761433 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-scripts" (OuterVolumeSpecName: "scripts") pod "5033e1a8-ab77-4894-ac5b-95ca39e731fa" (UID: "5033e1a8-ab77-4894-ac5b-95ca39e731fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.792540 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5033e1a8-ab77-4894-ac5b-95ca39e731fa" (UID: "5033e1a8-ab77-4894-ac5b-95ca39e731fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.807957 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-config-data" (OuterVolumeSpecName: "config-data") pod "5033e1a8-ab77-4894-ac5b-95ca39e731fa" (UID: "5033e1a8-ab77-4894-ac5b-95ca39e731fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.857321 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.857432 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.857450 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78zn8\" (UniqueName: \"kubernetes.io/projected/5033e1a8-ab77-4894-ac5b-95ca39e731fa-kube-api-access-78zn8\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:28 crc kubenswrapper[4726]: I1001 16:20:28.857460 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5033e1a8-ab77-4894-ac5b-95ca39e731fa-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:29 crc kubenswrapper[4726]: I1001 16:20:29.237870 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vj74b" event={"ID":"5033e1a8-ab77-4894-ac5b-95ca39e731fa","Type":"ContainerDied","Data":"d963a69db3db118290eb496b50bcafe982e4403fa72433d981c2e94721cdf69f"} Oct 01 16:20:29 crc kubenswrapper[4726]: I1001 16:20:29.238409 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d963a69db3db118290eb496b50bcafe982e4403fa72433d981c2e94721cdf69f" Oct 01 16:20:29 crc kubenswrapper[4726]: I1001 16:20:29.237965 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vj74b" Oct 01 16:20:29 crc kubenswrapper[4726]: I1001 16:20:29.426674 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:20:29 crc kubenswrapper[4726]: I1001 16:20:29.426933 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8000eb48-af65-45e9-abe6-3263671bc050" containerName="nova-api-log" containerID="cri-o://e586999f7364dfff65f561c4b34b9b526d9f937fa6053ef6a56d0d73904815a1" gracePeriod=30 Oct 01 16:20:29 crc kubenswrapper[4726]: I1001 16:20:29.427108 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8000eb48-af65-45e9-abe6-3263671bc050" containerName="nova-api-api" containerID="cri-o://030a821c9f34415d3a9c8a9e410bfc407619d81d38dcc9cb1e7a5422ece5b681" gracePeriod=30 Oct 01 16:20:29 crc kubenswrapper[4726]: I1001 16:20:29.463766 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:20:29 crc kubenswrapper[4726]: I1001 16:20:29.465152 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7d6747b1-097c-4c78-ad2f-1f2575dd5e46" containerName="nova-scheduler-scheduler" containerID="cri-o://e21317068070fadc0b29bd39c43364f8a1cdeaa244507389dbbf6e4f20858276" gracePeriod=30 Oct 01 16:20:29 crc kubenswrapper[4726]: I1001 16:20:29.482190 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:20:29 crc kubenswrapper[4726]: I1001 16:20:29.482411 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" containerName="nova-metadata-log" containerID="cri-o://d902b6ac6bfcf6d4a6342d6c60c1495659de816e5933d0a107c9dbf9af46fc98" gracePeriod=30 Oct 01 16:20:29 crc kubenswrapper[4726]: I1001 16:20:29.482704 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" containerName="nova-metadata-metadata" containerID="cri-o://bd147ac73eff6c440ece5ffed7152fa1c5387f469aaf8b342ffd9a6cbc1b294e" gracePeriod=30 Oct 01 16:20:30 crc kubenswrapper[4726]: I1001 16:20:30.254971 4726 generic.go:334] "Generic (PLEG): container finished" podID="8000eb48-af65-45e9-abe6-3263671bc050" containerID="e586999f7364dfff65f561c4b34b9b526d9f937fa6053ef6a56d0d73904815a1" exitCode=143 Oct 01 16:20:30 crc kubenswrapper[4726]: I1001 16:20:30.255077 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8000eb48-af65-45e9-abe6-3263671bc050","Type":"ContainerDied","Data":"e586999f7364dfff65f561c4b34b9b526d9f937fa6053ef6a56d0d73904815a1"} Oct 01 16:20:30 crc kubenswrapper[4726]: I1001 16:20:30.259160 4726 generic.go:334] "Generic (PLEG): container finished" podID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" containerID="d902b6ac6bfcf6d4a6342d6c60c1495659de816e5933d0a107c9dbf9af46fc98" exitCode=143 Oct 01 16:20:30 crc kubenswrapper[4726]: I1001 16:20:30.259221 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56e4e1ae-a938-4cf3-a34b-654dbcebeb51","Type":"ContainerDied","Data":"d902b6ac6bfcf6d4a6342d6c60c1495659de816e5933d0a107c9dbf9af46fc98"} Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.095772 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.206397 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-combined-ca-bundle\") pod \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\" (UID: \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\") " Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.206901 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-config-data\") pod \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\" (UID: \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\") " Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.207182 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gbrd\" (UniqueName: \"kubernetes.io/projected/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-kube-api-access-7gbrd\") pod \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\" (UID: \"7d6747b1-097c-4c78-ad2f-1f2575dd5e46\") " Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.213399 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-kube-api-access-7gbrd" (OuterVolumeSpecName: "kube-api-access-7gbrd") pod "7d6747b1-097c-4c78-ad2f-1f2575dd5e46" (UID: "7d6747b1-097c-4c78-ad2f-1f2575dd5e46"). InnerVolumeSpecName "kube-api-access-7gbrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.246045 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-config-data" (OuterVolumeSpecName: "config-data") pod "7d6747b1-097c-4c78-ad2f-1f2575dd5e46" (UID: "7d6747b1-097c-4c78-ad2f-1f2575dd5e46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.263933 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d6747b1-097c-4c78-ad2f-1f2575dd5e46" (UID: "7d6747b1-097c-4c78-ad2f-1f2575dd5e46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.271119 4726 generic.go:334] "Generic (PLEG): container finished" podID="7d6747b1-097c-4c78-ad2f-1f2575dd5e46" containerID="e21317068070fadc0b29bd39c43364f8a1cdeaa244507389dbbf6e4f20858276" exitCode=0 Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.271166 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7d6747b1-097c-4c78-ad2f-1f2575dd5e46","Type":"ContainerDied","Data":"e21317068070fadc0b29bd39c43364f8a1cdeaa244507389dbbf6e4f20858276"} Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.271180 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.271195 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7d6747b1-097c-4c78-ad2f-1f2575dd5e46","Type":"ContainerDied","Data":"f79afda308a18c7b4caeace34a93afd6e5f005d4b8e869b9731bd378b660a29b"} Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.271215 4726 scope.go:117] "RemoveContainer" containerID="e21317068070fadc0b29bd39c43364f8a1cdeaa244507389dbbf6e4f20858276" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.310017 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.310099 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gbrd\" (UniqueName: \"kubernetes.io/projected/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-kube-api-access-7gbrd\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.310113 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d6747b1-097c-4c78-ad2f-1f2575dd5e46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.351329 4726 scope.go:117] "RemoveContainer" containerID="e21317068070fadc0b29bd39c43364f8a1cdeaa244507389dbbf6e4f20858276" Oct 01 16:20:31 crc kubenswrapper[4726]: E1001 16:20:31.351854 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e21317068070fadc0b29bd39c43364f8a1cdeaa244507389dbbf6e4f20858276\": container with ID starting with e21317068070fadc0b29bd39c43364f8a1cdeaa244507389dbbf6e4f20858276 not found: ID does not exist" containerID="e21317068070fadc0b29bd39c43364f8a1cdeaa244507389dbbf6e4f20858276" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.351946 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e21317068070fadc0b29bd39c43364f8a1cdeaa244507389dbbf6e4f20858276"} err="failed to get container status \"e21317068070fadc0b29bd39c43364f8a1cdeaa244507389dbbf6e4f20858276\": rpc error: code = NotFound desc = could not find container \"e21317068070fadc0b29bd39c43364f8a1cdeaa244507389dbbf6e4f20858276\": container with ID starting with e21317068070fadc0b29bd39c43364f8a1cdeaa244507389dbbf6e4f20858276 not found: ID does not exist" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.371254 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.379929 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.390962 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:20:31 crc kubenswrapper[4726]: E1001 16:20:31.391546 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" containerName="init" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.391611 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" containerName="init" Oct 01 16:20:31 crc kubenswrapper[4726]: E1001 16:20:31.391682 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5033e1a8-ab77-4894-ac5b-95ca39e731fa" containerName="nova-manage" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.391730 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5033e1a8-ab77-4894-ac5b-95ca39e731fa" containerName="nova-manage" Oct 01 16:20:31 crc kubenswrapper[4726]: E1001 16:20:31.391785 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" containerName="dnsmasq-dns" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.391830 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" containerName="dnsmasq-dns" Oct 01 16:20:31 crc kubenswrapper[4726]: E1001 16:20:31.391894 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d6747b1-097c-4c78-ad2f-1f2575dd5e46" containerName="nova-scheduler-scheduler" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.391941 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d6747b1-097c-4c78-ad2f-1f2575dd5e46" containerName="nova-scheduler-scheduler" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.392180 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="60e46f07-6ffb-46ca-8d0d-a5749dfd42f6" containerName="dnsmasq-dns" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.392251 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5033e1a8-ab77-4894-ac5b-95ca39e731fa" containerName="nova-manage" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.392306 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d6747b1-097c-4c78-ad2f-1f2575dd5e46" containerName="nova-scheduler-scheduler" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.392981 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.395535 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.399505 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.513380 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c094983e-ab3f-4359-a0ed-5481f349409a-config-data\") pod \"nova-scheduler-0\" (UID: \"c094983e-ab3f-4359-a0ed-5481f349409a\") " pod="openstack/nova-scheduler-0" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.513492 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c094983e-ab3f-4359-a0ed-5481f349409a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c094983e-ab3f-4359-a0ed-5481f349409a\") " pod="openstack/nova-scheduler-0" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.513539 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9jjt\" (UniqueName: \"kubernetes.io/projected/c094983e-ab3f-4359-a0ed-5481f349409a-kube-api-access-h9jjt\") pod \"nova-scheduler-0\" (UID: \"c094983e-ab3f-4359-a0ed-5481f349409a\") " pod="openstack/nova-scheduler-0" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.616281 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c094983e-ab3f-4359-a0ed-5481f349409a-config-data\") pod \"nova-scheduler-0\" (UID: \"c094983e-ab3f-4359-a0ed-5481f349409a\") " pod="openstack/nova-scheduler-0" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.616429 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c094983e-ab3f-4359-a0ed-5481f349409a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c094983e-ab3f-4359-a0ed-5481f349409a\") " pod="openstack/nova-scheduler-0" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.616497 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9jjt\" (UniqueName: \"kubernetes.io/projected/c094983e-ab3f-4359-a0ed-5481f349409a-kube-api-access-h9jjt\") pod \"nova-scheduler-0\" (UID: \"c094983e-ab3f-4359-a0ed-5481f349409a\") " pod="openstack/nova-scheduler-0" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.620335 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c094983e-ab3f-4359-a0ed-5481f349409a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c094983e-ab3f-4359-a0ed-5481f349409a\") " pod="openstack/nova-scheduler-0" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.620424 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c094983e-ab3f-4359-a0ed-5481f349409a-config-data\") pod \"nova-scheduler-0\" (UID: \"c094983e-ab3f-4359-a0ed-5481f349409a\") " pod="openstack/nova-scheduler-0" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.641794 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9jjt\" (UniqueName: \"kubernetes.io/projected/c094983e-ab3f-4359-a0ed-5481f349409a-kube-api-access-h9jjt\") pod \"nova-scheduler-0\" (UID: \"c094983e-ab3f-4359-a0ed-5481f349409a\") " pod="openstack/nova-scheduler-0" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.712321 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:20:31 crc kubenswrapper[4726]: I1001 16:20:31.819148 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d6747b1-097c-4c78-ad2f-1f2575dd5e46" path="/var/lib/kubelet/pods/7d6747b1-097c-4c78-ad2f-1f2575dd5e46/volumes" Oct 01 16:20:32 crc kubenswrapper[4726]: I1001 16:20:32.171615 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:20:32 crc kubenswrapper[4726]: I1001 16:20:32.282120 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c094983e-ab3f-4359-a0ed-5481f349409a","Type":"ContainerStarted","Data":"54acb2c49be12b8dcb0f938f67c654894398ed15cae4006e4feb31957d971438"} Oct 01 16:20:32 crc kubenswrapper[4726]: I1001 16:20:32.610959 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:41464->10.217.0.194:8775: read: connection reset by peer" Oct 01 16:20:32 crc kubenswrapper[4726]: I1001 16:20:32.610996 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:41458->10.217.0.194:8775: read: connection reset by peer" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.123236 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.249654 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-nova-metadata-tls-certs\") pod \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.249884 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-combined-ca-bundle\") pod \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.249948 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-logs\") pod \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.250035 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-config-data\") pod \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.250199 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnmn5\" (UniqueName: \"kubernetes.io/projected/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-kube-api-access-nnmn5\") pod \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\" (UID: \"56e4e1ae-a938-4cf3-a34b-654dbcebeb51\") " Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.251138 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-logs" (OuterVolumeSpecName: "logs") pod "56e4e1ae-a938-4cf3-a34b-654dbcebeb51" (UID: "56e4e1ae-a938-4cf3-a34b-654dbcebeb51"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.255031 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-kube-api-access-nnmn5" (OuterVolumeSpecName: "kube-api-access-nnmn5") pod "56e4e1ae-a938-4cf3-a34b-654dbcebeb51" (UID: "56e4e1ae-a938-4cf3-a34b-654dbcebeb51"). InnerVolumeSpecName "kube-api-access-nnmn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.277291 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56e4e1ae-a938-4cf3-a34b-654dbcebeb51" (UID: "56e4e1ae-a938-4cf3-a34b-654dbcebeb51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.287481 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-config-data" (OuterVolumeSpecName: "config-data") pod "56e4e1ae-a938-4cf3-a34b-654dbcebeb51" (UID: "56e4e1ae-a938-4cf3-a34b-654dbcebeb51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.297986 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c094983e-ab3f-4359-a0ed-5481f349409a","Type":"ContainerStarted","Data":"05ac8c38d4f21aa83ce148294ef77052c2b4e0d8d24f0fe217f5b6071b10c146"} Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.304309 4726 generic.go:334] "Generic (PLEG): container finished" podID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" containerID="bd147ac73eff6c440ece5ffed7152fa1c5387f469aaf8b342ffd9a6cbc1b294e" exitCode=0 Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.304363 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.304516 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56e4e1ae-a938-4cf3-a34b-654dbcebeb51","Type":"ContainerDied","Data":"bd147ac73eff6c440ece5ffed7152fa1c5387f469aaf8b342ffd9a6cbc1b294e"} Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.304630 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56e4e1ae-a938-4cf3-a34b-654dbcebeb51","Type":"ContainerDied","Data":"e23d6ff9abec3c388c3fe07f77eb01ee3aece7f51ffa7d70ca929256fd2dfc14"} Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.304732 4726 scope.go:117] "RemoveContainer" containerID="bd147ac73eff6c440ece5ffed7152fa1c5387f469aaf8b342ffd9a6cbc1b294e" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.315538 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "56e4e1ae-a938-4cf3-a34b-654dbcebeb51" (UID: "56e4e1ae-a938-4cf3-a34b-654dbcebeb51"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.324596 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.324569076 podStartE2EDuration="2.324569076s" podCreationTimestamp="2025-10-01 16:20:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:20:33.316098048 +0000 UTC m=+1286.217650635" watchObservedRunningTime="2025-10-01 16:20:33.324569076 +0000 UTC m=+1286.226121653" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.352272 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.352313 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.352327 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.352349 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnmn5\" (UniqueName: \"kubernetes.io/projected/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-kube-api-access-nnmn5\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.352364 4726 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e4e1ae-a938-4cf3-a34b-654dbcebeb51-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.394330 4726 scope.go:117] "RemoveContainer" containerID="d902b6ac6bfcf6d4a6342d6c60c1495659de816e5933d0a107c9dbf9af46fc98" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.417110 4726 scope.go:117] "RemoveContainer" containerID="bd147ac73eff6c440ece5ffed7152fa1c5387f469aaf8b342ffd9a6cbc1b294e" Oct 01 16:20:33 crc kubenswrapper[4726]: E1001 16:20:33.417742 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd147ac73eff6c440ece5ffed7152fa1c5387f469aaf8b342ffd9a6cbc1b294e\": container with ID starting with bd147ac73eff6c440ece5ffed7152fa1c5387f469aaf8b342ffd9a6cbc1b294e not found: ID does not exist" containerID="bd147ac73eff6c440ece5ffed7152fa1c5387f469aaf8b342ffd9a6cbc1b294e" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.417873 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd147ac73eff6c440ece5ffed7152fa1c5387f469aaf8b342ffd9a6cbc1b294e"} err="failed to get container status \"bd147ac73eff6c440ece5ffed7152fa1c5387f469aaf8b342ffd9a6cbc1b294e\": rpc error: code = NotFound desc = could not find container \"bd147ac73eff6c440ece5ffed7152fa1c5387f469aaf8b342ffd9a6cbc1b294e\": container with ID starting with bd147ac73eff6c440ece5ffed7152fa1c5387f469aaf8b342ffd9a6cbc1b294e not found: ID does not exist" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.417967 4726 scope.go:117] "RemoveContainer" containerID="d902b6ac6bfcf6d4a6342d6c60c1495659de816e5933d0a107c9dbf9af46fc98" Oct 01 16:20:33 crc kubenswrapper[4726]: E1001 16:20:33.418343 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d902b6ac6bfcf6d4a6342d6c60c1495659de816e5933d0a107c9dbf9af46fc98\": container with ID starting with d902b6ac6bfcf6d4a6342d6c60c1495659de816e5933d0a107c9dbf9af46fc98 not found: ID does not exist" containerID="d902b6ac6bfcf6d4a6342d6c60c1495659de816e5933d0a107c9dbf9af46fc98" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.418444 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d902b6ac6bfcf6d4a6342d6c60c1495659de816e5933d0a107c9dbf9af46fc98"} err="failed to get container status \"d902b6ac6bfcf6d4a6342d6c60c1495659de816e5933d0a107c9dbf9af46fc98\": rpc error: code = NotFound desc = could not find container \"d902b6ac6bfcf6d4a6342d6c60c1495659de816e5933d0a107c9dbf9af46fc98\": container with ID starting with d902b6ac6bfcf6d4a6342d6c60c1495659de816e5933d0a107c9dbf9af46fc98 not found: ID does not exist" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.645243 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.659931 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.687295 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:20:33 crc kubenswrapper[4726]: E1001 16:20:33.687954 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" containerName="nova-metadata-metadata" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.687969 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" containerName="nova-metadata-metadata" Oct 01 16:20:33 crc kubenswrapper[4726]: E1001 16:20:33.688008 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" containerName="nova-metadata-log" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.688015 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" containerName="nova-metadata-log" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.688979 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" containerName="nova-metadata-metadata" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.689014 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" containerName="nova-metadata-log" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.691413 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.694803 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.695010 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.695851 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.819204 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56e4e1ae-a938-4cf3-a34b-654dbcebeb51" path="/var/lib/kubelet/pods/56e4e1ae-a938-4cf3-a34b-654dbcebeb51/volumes" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.861002 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-logs\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.861096 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfl4v\" (UniqueName: \"kubernetes.io/projected/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-kube-api-access-pfl4v\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.861133 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.861201 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.861290 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-config-data\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.963038 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.963370 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-config-data\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.963509 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-logs\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.963635 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfl4v\" (UniqueName: \"kubernetes.io/projected/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-kube-api-access-pfl4v\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.963733 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.964818 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-logs\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.968219 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.969215 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.969388 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-config-data\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:33 crc kubenswrapper[4726]: I1001 16:20:33.981615 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfl4v\" (UniqueName: \"kubernetes.io/projected/8ec3f7f2-1bb9-47a0-9f62-54625de5bf67-kube-api-access-pfl4v\") pod \"nova-metadata-0\" (UID: \"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67\") " pod="openstack/nova-metadata-0" Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.015142 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.317105 4726 generic.go:334] "Generic (PLEG): container finished" podID="8000eb48-af65-45e9-abe6-3263671bc050" containerID="030a821c9f34415d3a9c8a9e410bfc407619d81d38dcc9cb1e7a5422ece5b681" exitCode=0 Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.317191 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8000eb48-af65-45e9-abe6-3263671bc050","Type":"ContainerDied","Data":"030a821c9f34415d3a9c8a9e410bfc407619d81d38dcc9cb1e7a5422ece5b681"} Oct 01 16:20:34 crc kubenswrapper[4726]: W1001 16:20:34.519877 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ec3f7f2_1bb9_47a0_9f62_54625de5bf67.slice/crio-5e87fb6fdf52dd6121f406b0ab3136b111028f5b69f4c0c23d86a2a549e8fab8 WatchSource:0}: Error finding container 5e87fb6fdf52dd6121f406b0ab3136b111028f5b69f4c0c23d86a2a549e8fab8: Status 404 returned error can't find the container with id 5e87fb6fdf52dd6121f406b0ab3136b111028f5b69f4c0c23d86a2a549e8fab8 Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.521200 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.732465 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.889674 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-internal-tls-certs\") pod \"8000eb48-af65-45e9-abe6-3263671bc050\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.889781 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-combined-ca-bundle\") pod \"8000eb48-af65-45e9-abe6-3263671bc050\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.889851 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8vcx\" (UniqueName: \"kubernetes.io/projected/8000eb48-af65-45e9-abe6-3263671bc050-kube-api-access-j8vcx\") pod \"8000eb48-af65-45e9-abe6-3263671bc050\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.889926 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-public-tls-certs\") pod \"8000eb48-af65-45e9-abe6-3263671bc050\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.889994 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-config-data\") pod \"8000eb48-af65-45e9-abe6-3263671bc050\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.890113 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8000eb48-af65-45e9-abe6-3263671bc050-logs\") pod \"8000eb48-af65-45e9-abe6-3263671bc050\" (UID: \"8000eb48-af65-45e9-abe6-3263671bc050\") " Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.890595 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8000eb48-af65-45e9-abe6-3263671bc050-logs" (OuterVolumeSpecName: "logs") pod "8000eb48-af65-45e9-abe6-3263671bc050" (UID: "8000eb48-af65-45e9-abe6-3263671bc050"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.891108 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8000eb48-af65-45e9-abe6-3263671bc050-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.896169 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8000eb48-af65-45e9-abe6-3263671bc050-kube-api-access-j8vcx" (OuterVolumeSpecName: "kube-api-access-j8vcx") pod "8000eb48-af65-45e9-abe6-3263671bc050" (UID: "8000eb48-af65-45e9-abe6-3263671bc050"). InnerVolumeSpecName "kube-api-access-j8vcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.915786 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-config-data" (OuterVolumeSpecName: "config-data") pod "8000eb48-af65-45e9-abe6-3263671bc050" (UID: "8000eb48-af65-45e9-abe6-3263671bc050"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.920315 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8000eb48-af65-45e9-abe6-3263671bc050" (UID: "8000eb48-af65-45e9-abe6-3263671bc050"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.943618 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8000eb48-af65-45e9-abe6-3263671bc050" (UID: "8000eb48-af65-45e9-abe6-3263671bc050"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.955849 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8000eb48-af65-45e9-abe6-3263671bc050" (UID: "8000eb48-af65-45e9-abe6-3263671bc050"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.992788 4726 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.992822 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.992832 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8vcx\" (UniqueName: \"kubernetes.io/projected/8000eb48-af65-45e9-abe6-3263671bc050-kube-api-access-j8vcx\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.992842 4726 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:34 crc kubenswrapper[4726]: I1001 16:20:34.992851 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8000eb48-af65-45e9-abe6-3263671bc050-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.331141 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67","Type":"ContainerStarted","Data":"50346e4ada4e2d6e8e7779336942b2418043e92debc5ab8121818eaa31772490"} Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.331515 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67","Type":"ContainerStarted","Data":"42a96b36a87539446926fa8b5a0f5930d4a18f98ebc57caeb80ce9d2a5e54be3"} Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.331531 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8ec3f7f2-1bb9-47a0-9f62-54625de5bf67","Type":"ContainerStarted","Data":"5e87fb6fdf52dd6121f406b0ab3136b111028f5b69f4c0c23d86a2a549e8fab8"} Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.333236 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8000eb48-af65-45e9-abe6-3263671bc050","Type":"ContainerDied","Data":"22254b36c44abe58270a2a399e0c47dc416cb63506be13393ac11f01b692cf9c"} Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.333295 4726 scope.go:117] "RemoveContainer" containerID="030a821c9f34415d3a9c8a9e410bfc407619d81d38dcc9cb1e7a5422ece5b681" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.333305 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.361580 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.361556384 podStartE2EDuration="2.361556384s" podCreationTimestamp="2025-10-01 16:20:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:20:35.351563683 +0000 UTC m=+1288.253116260" watchObservedRunningTime="2025-10-01 16:20:35.361556384 +0000 UTC m=+1288.263108981" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.382827 4726 scope.go:117] "RemoveContainer" containerID="e586999f7364dfff65f561c4b34b9b526d9f937fa6053ef6a56d0d73904815a1" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.403583 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.417596 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.433300 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 16:20:35 crc kubenswrapper[4726]: E1001 16:20:35.433849 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8000eb48-af65-45e9-abe6-3263671bc050" containerName="nova-api-log" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.433871 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8000eb48-af65-45e9-abe6-3263671bc050" containerName="nova-api-log" Oct 01 16:20:35 crc kubenswrapper[4726]: E1001 16:20:35.433895 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8000eb48-af65-45e9-abe6-3263671bc050" containerName="nova-api-api" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.433904 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8000eb48-af65-45e9-abe6-3263671bc050" containerName="nova-api-api" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.434141 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8000eb48-af65-45e9-abe6-3263671bc050" containerName="nova-api-api" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.434164 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8000eb48-af65-45e9-abe6-3263671bc050" containerName="nova-api-log" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.435392 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.438403 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.438490 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.438798 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.445719 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.603979 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6188ef34-0980-4918-bb75-27886226804e-logs\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.604033 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49zb2\" (UniqueName: \"kubernetes.io/projected/6188ef34-0980-4918-bb75-27886226804e-kube-api-access-49zb2\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.604080 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6188ef34-0980-4918-bb75-27886226804e-config-data\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.604133 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6188ef34-0980-4918-bb75-27886226804e-public-tls-certs\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.604161 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6188ef34-0980-4918-bb75-27886226804e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.604179 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6188ef34-0980-4918-bb75-27886226804e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.712163 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6188ef34-0980-4918-bb75-27886226804e-logs\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.712303 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49zb2\" (UniqueName: \"kubernetes.io/projected/6188ef34-0980-4918-bb75-27886226804e-kube-api-access-49zb2\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.712376 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6188ef34-0980-4918-bb75-27886226804e-config-data\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.712495 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6188ef34-0980-4918-bb75-27886226804e-public-tls-certs\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.712552 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6188ef34-0980-4918-bb75-27886226804e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.713010 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6188ef34-0980-4918-bb75-27886226804e-logs\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.713918 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6188ef34-0980-4918-bb75-27886226804e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.722168 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6188ef34-0980-4918-bb75-27886226804e-public-tls-certs\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.722871 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6188ef34-0980-4918-bb75-27886226804e-config-data\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.722872 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6188ef34-0980-4918-bb75-27886226804e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.729783 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6188ef34-0980-4918-bb75-27886226804e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.745019 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49zb2\" (UniqueName: \"kubernetes.io/projected/6188ef34-0980-4918-bb75-27886226804e-kube-api-access-49zb2\") pod \"nova-api-0\" (UID: \"6188ef34-0980-4918-bb75-27886226804e\") " pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.802849 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:20:35 crc kubenswrapper[4726]: I1001 16:20:35.830321 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8000eb48-af65-45e9-abe6-3263671bc050" path="/var/lib/kubelet/pods/8000eb48-af65-45e9-abe6-3263671bc050/volumes" Oct 01 16:20:36 crc kubenswrapper[4726]: I1001 16:20:36.273285 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:20:36 crc kubenswrapper[4726]: W1001 16:20:36.282605 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6188ef34_0980_4918_bb75_27886226804e.slice/crio-55c18a7a96dc03b8996bcf52599a33393fa248b9f2201188b6dc1700796ab92e WatchSource:0}: Error finding container 55c18a7a96dc03b8996bcf52599a33393fa248b9f2201188b6dc1700796ab92e: Status 404 returned error can't find the container with id 55c18a7a96dc03b8996bcf52599a33393fa248b9f2201188b6dc1700796ab92e Oct 01 16:20:36 crc kubenswrapper[4726]: I1001 16:20:36.348575 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6188ef34-0980-4918-bb75-27886226804e","Type":"ContainerStarted","Data":"55c18a7a96dc03b8996bcf52599a33393fa248b9f2201188b6dc1700796ab92e"} Oct 01 16:20:36 crc kubenswrapper[4726]: I1001 16:20:36.712983 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 16:20:37 crc kubenswrapper[4726]: I1001 16:20:37.361667 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6188ef34-0980-4918-bb75-27886226804e","Type":"ContainerStarted","Data":"ee48a8dc81e2e040ae9fa85236222ce31f8d407eaf5bd27a15190ad55af11474"} Oct 01 16:20:37 crc kubenswrapper[4726]: I1001 16:20:37.361714 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6188ef34-0980-4918-bb75-27886226804e","Type":"ContainerStarted","Data":"3590fb7a4b0e03ed5954b30677000faa1ff781140a7e0f775c5e44327bb6cc57"} Oct 01 16:20:39 crc kubenswrapper[4726]: I1001 16:20:39.015651 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 16:20:39 crc kubenswrapper[4726]: I1001 16:20:39.016189 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 16:20:41 crc kubenswrapper[4726]: I1001 16:20:41.712628 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 01 16:20:41 crc kubenswrapper[4726]: I1001 16:20:41.748400 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 01 16:20:41 crc kubenswrapper[4726]: I1001 16:20:41.774006 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=6.773988546 podStartE2EDuration="6.773988546s" podCreationTimestamp="2025-10-01 16:20:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:20:37.384396945 +0000 UTC m=+1290.285949522" watchObservedRunningTime="2025-10-01 16:20:41.773988546 +0000 UTC m=+1294.675541133" Oct 01 16:20:42 crc kubenswrapper[4726]: I1001 16:20:42.457371 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 01 16:20:44 crc kubenswrapper[4726]: I1001 16:20:44.015984 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 16:20:44 crc kubenswrapper[4726]: I1001 16:20:44.016361 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 16:20:45 crc kubenswrapper[4726]: I1001 16:20:45.030334 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8ec3f7f2-1bb9-47a0-9f62-54625de5bf67" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:20:45 crc kubenswrapper[4726]: I1001 16:20:45.030365 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8ec3f7f2-1bb9-47a0-9f62-54625de5bf67" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:20:45 crc kubenswrapper[4726]: I1001 16:20:45.803494 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:20:45 crc kubenswrapper[4726]: I1001 16:20:45.804420 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:20:46 crc kubenswrapper[4726]: I1001 16:20:46.818284 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6188ef34-0980-4918-bb75-27886226804e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:20:46 crc kubenswrapper[4726]: I1001 16:20:46.818658 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6188ef34-0980-4918-bb75-27886226804e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:20:47 crc kubenswrapper[4726]: I1001 16:20:47.887202 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 01 16:20:53 crc kubenswrapper[4726]: I1001 16:20:53.414001 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:20:53 crc kubenswrapper[4726]: I1001 16:20:53.415026 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:20:54 crc kubenswrapper[4726]: I1001 16:20:54.021639 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 16:20:54 crc kubenswrapper[4726]: I1001 16:20:54.022527 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 16:20:54 crc kubenswrapper[4726]: I1001 16:20:54.028816 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 16:20:54 crc kubenswrapper[4726]: I1001 16:20:54.552235 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 16:20:55 crc kubenswrapper[4726]: I1001 16:20:55.880952 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 16:20:55 crc kubenswrapper[4726]: I1001 16:20:55.881421 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 16:20:55 crc kubenswrapper[4726]: I1001 16:20:55.881991 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 16:20:55 crc kubenswrapper[4726]: I1001 16:20:55.882007 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 16:20:55 crc kubenswrapper[4726]: I1001 16:20:55.890186 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 16:20:55 crc kubenswrapper[4726]: I1001 16:20:55.894559 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 16:21:04 crc kubenswrapper[4726]: I1001 16:21:04.359285 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:21:05 crc kubenswrapper[4726]: I1001 16:21:05.146106 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:21:08 crc kubenswrapper[4726]: I1001 16:21:08.523350 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="57593043-8bb5-4967-8a6c-1a55a3e8894c" containerName="rabbitmq" containerID="cri-o://8217cab599a9ea7e8ecc95c31f47ebced3aad20e93476a47c5964c4cb700cbf8" gracePeriod=604796 Oct 01 16:21:09 crc kubenswrapper[4726]: I1001 16:21:09.903777 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="75dbebbf-c4f3-4508-901d-05148fd48f74" containerName="rabbitmq" containerID="cri-o://cb8f6b26ed1180f41f0273e2461ce4276f12488072d6c566ce51a960f48e6ec2" gracePeriod=604796 Oct 01 16:21:14 crc kubenswrapper[4726]: I1001 16:21:14.789894 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="57593043-8bb5-4967-8a6c-1a55a3e8894c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Oct 01 16:21:14 crc kubenswrapper[4726]: I1001 16:21:14.836399 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="75dbebbf-c4f3-4508-901d-05148fd48f74" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.226484 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.296806 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"57593043-8bb5-4967-8a6c-1a55a3e8894c\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.297032 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-config-data\") pod \"57593043-8bb5-4967-8a6c-1a55a3e8894c\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.297101 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-plugins\") pod \"57593043-8bb5-4967-8a6c-1a55a3e8894c\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.297139 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/57593043-8bb5-4967-8a6c-1a55a3e8894c-pod-info\") pod \"57593043-8bb5-4967-8a6c-1a55a3e8894c\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.297199 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-confd\") pod \"57593043-8bb5-4967-8a6c-1a55a3e8894c\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.297227 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-tls\") pod \"57593043-8bb5-4967-8a6c-1a55a3e8894c\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.297280 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qf5k\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-kube-api-access-6qf5k\") pod \"57593043-8bb5-4967-8a6c-1a55a3e8894c\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.298943 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "57593043-8bb5-4967-8a6c-1a55a3e8894c" (UID: "57593043-8bb5-4967-8a6c-1a55a3e8894c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.304023 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/57593043-8bb5-4967-8a6c-1a55a3e8894c-pod-info" (OuterVolumeSpecName: "pod-info") pod "57593043-8bb5-4967-8a6c-1a55a3e8894c" (UID: "57593043-8bb5-4967-8a6c-1a55a3e8894c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.304415 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "57593043-8bb5-4967-8a6c-1a55a3e8894c" (UID: "57593043-8bb5-4967-8a6c-1a55a3e8894c"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.304650 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-kube-api-access-6qf5k" (OuterVolumeSpecName: "kube-api-access-6qf5k") pod "57593043-8bb5-4967-8a6c-1a55a3e8894c" (UID: "57593043-8bb5-4967-8a6c-1a55a3e8894c"). InnerVolumeSpecName "kube-api-access-6qf5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.316212 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "57593043-8bb5-4967-8a6c-1a55a3e8894c" (UID: "57593043-8bb5-4967-8a6c-1a55a3e8894c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.354915 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-config-data" (OuterVolumeSpecName: "config-data") pod "57593043-8bb5-4967-8a6c-1a55a3e8894c" (UID: "57593043-8bb5-4967-8a6c-1a55a3e8894c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.398961 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/57593043-8bb5-4967-8a6c-1a55a3e8894c-erlang-cookie-secret\") pod \"57593043-8bb5-4967-8a6c-1a55a3e8894c\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.399061 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-erlang-cookie\") pod \"57593043-8bb5-4967-8a6c-1a55a3e8894c\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.399146 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-plugins-conf\") pod \"57593043-8bb5-4967-8a6c-1a55a3e8894c\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.399314 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-server-conf\") pod \"57593043-8bb5-4967-8a6c-1a55a3e8894c\" (UID: \"57593043-8bb5-4967-8a6c-1a55a3e8894c\") " Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.399720 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.399739 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.399748 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.399757 4726 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/57593043-8bb5-4967-8a6c-1a55a3e8894c-pod-info\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.399766 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.399776 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qf5k\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-kube-api-access-6qf5k\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.399778 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "57593043-8bb5-4967-8a6c-1a55a3e8894c" (UID: "57593043-8bb5-4967-8a6c-1a55a3e8894c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.399964 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "57593043-8bb5-4967-8a6c-1a55a3e8894c" (UID: "57593043-8bb5-4967-8a6c-1a55a3e8894c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.402726 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57593043-8bb5-4967-8a6c-1a55a3e8894c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "57593043-8bb5-4967-8a6c-1a55a3e8894c" (UID: "57593043-8bb5-4967-8a6c-1a55a3e8894c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.422539 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.453463 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "57593043-8bb5-4967-8a6c-1a55a3e8894c" (UID: "57593043-8bb5-4967-8a6c-1a55a3e8894c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.457396 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-server-conf" (OuterVolumeSpecName: "server-conf") pod "57593043-8bb5-4967-8a6c-1a55a3e8894c" (UID: "57593043-8bb5-4967-8a6c-1a55a3e8894c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.501586 4726 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.501618 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.501630 4726 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/57593043-8bb5-4967-8a6c-1a55a3e8894c-server-conf\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.501639 4726 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/57593043-8bb5-4967-8a6c-1a55a3e8894c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.501648 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.501657 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/57593043-8bb5-4967-8a6c-1a55a3e8894c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.800414 4726 generic.go:334] "Generic (PLEG): container finished" podID="57593043-8bb5-4967-8a6c-1a55a3e8894c" containerID="8217cab599a9ea7e8ecc95c31f47ebced3aad20e93476a47c5964c4cb700cbf8" exitCode=0 Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.800466 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"57593043-8bb5-4967-8a6c-1a55a3e8894c","Type":"ContainerDied","Data":"8217cab599a9ea7e8ecc95c31f47ebced3aad20e93476a47c5964c4cb700cbf8"} Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.800503 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"57593043-8bb5-4967-8a6c-1a55a3e8894c","Type":"ContainerDied","Data":"5080ba6f47df3de6a066783f373eaa1014b63a486e183c8bf31f9c881e568b04"} Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.800523 4726 scope.go:117] "RemoveContainer" containerID="8217cab599a9ea7e8ecc95c31f47ebced3aad20e93476a47c5964c4cb700cbf8" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.800470 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.829026 4726 scope.go:117] "RemoveContainer" containerID="2deafad37316839f98e372793ca49f200759055adcb7e9e7e730bfd5d7b19193" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.849162 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.858822 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.887308 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:21:15 crc kubenswrapper[4726]: E1001 16:21:15.887663 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57593043-8bb5-4967-8a6c-1a55a3e8894c" containerName="rabbitmq" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.887678 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="57593043-8bb5-4967-8a6c-1a55a3e8894c" containerName="rabbitmq" Oct 01 16:21:15 crc kubenswrapper[4726]: E1001 16:21:15.887700 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57593043-8bb5-4967-8a6c-1a55a3e8894c" containerName="setup-container" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.887708 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="57593043-8bb5-4967-8a6c-1a55a3e8894c" containerName="setup-container" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.890897 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="57593043-8bb5-4967-8a6c-1a55a3e8894c" containerName="rabbitmq" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.892216 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.895139 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.895325 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.895902 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-b2sq4" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.896102 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.896224 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.896340 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.896439 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.905444 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.906663 4726 scope.go:117] "RemoveContainer" containerID="8217cab599a9ea7e8ecc95c31f47ebced3aad20e93476a47c5964c4cb700cbf8" Oct 01 16:21:15 crc kubenswrapper[4726]: E1001 16:21:15.908298 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8217cab599a9ea7e8ecc95c31f47ebced3aad20e93476a47c5964c4cb700cbf8\": container with ID starting with 8217cab599a9ea7e8ecc95c31f47ebced3aad20e93476a47c5964c4cb700cbf8 not found: ID does not exist" containerID="8217cab599a9ea7e8ecc95c31f47ebced3aad20e93476a47c5964c4cb700cbf8" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.908442 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8217cab599a9ea7e8ecc95c31f47ebced3aad20e93476a47c5964c4cb700cbf8"} err="failed to get container status \"8217cab599a9ea7e8ecc95c31f47ebced3aad20e93476a47c5964c4cb700cbf8\": rpc error: code = NotFound desc = could not find container \"8217cab599a9ea7e8ecc95c31f47ebced3aad20e93476a47c5964c4cb700cbf8\": container with ID starting with 8217cab599a9ea7e8ecc95c31f47ebced3aad20e93476a47c5964c4cb700cbf8 not found: ID does not exist" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.908521 4726 scope.go:117] "RemoveContainer" containerID="2deafad37316839f98e372793ca49f200759055adcb7e9e7e730bfd5d7b19193" Oct 01 16:21:15 crc kubenswrapper[4726]: E1001 16:21:15.911003 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2deafad37316839f98e372793ca49f200759055adcb7e9e7e730bfd5d7b19193\": container with ID starting with 2deafad37316839f98e372793ca49f200759055adcb7e9e7e730bfd5d7b19193 not found: ID does not exist" containerID="2deafad37316839f98e372793ca49f200759055adcb7e9e7e730bfd5d7b19193" Oct 01 16:21:15 crc kubenswrapper[4726]: I1001 16:21:15.911093 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2deafad37316839f98e372793ca49f200759055adcb7e9e7e730bfd5d7b19193"} err="failed to get container status \"2deafad37316839f98e372793ca49f200759055adcb7e9e7e730bfd5d7b19193\": rpc error: code = NotFound desc = could not find container \"2deafad37316839f98e372793ca49f200759055adcb7e9e7e730bfd5d7b19193\": container with ID starting with 2deafad37316839f98e372793ca49f200759055adcb7e9e7e730bfd5d7b19193 not found: ID does not exist" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.010744 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b5c68163-e0cf-43e8-8659-260fb0aca643-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.010847 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b5c68163-e0cf-43e8-8659-260fb0aca643-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.010875 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b5c68163-e0cf-43e8-8659-260fb0aca643-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.010903 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b5c68163-e0cf-43e8-8659-260fb0aca643-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.010922 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b5c68163-e0cf-43e8-8659-260fb0aca643-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.010972 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b5c68163-e0cf-43e8-8659-260fb0aca643-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.011167 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gpmf\" (UniqueName: \"kubernetes.io/projected/b5c68163-e0cf-43e8-8659-260fb0aca643-kube-api-access-5gpmf\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.011230 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b5c68163-e0cf-43e8-8659-260fb0aca643-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.011292 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.011565 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b5c68163-e0cf-43e8-8659-260fb0aca643-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.011644 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b5c68163-e0cf-43e8-8659-260fb0aca643-config-data\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.113193 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b5c68163-e0cf-43e8-8659-260fb0aca643-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.113472 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b5c68163-e0cf-43e8-8659-260fb0aca643-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.113591 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b5c68163-e0cf-43e8-8659-260fb0aca643-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.113689 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b5c68163-e0cf-43e8-8659-260fb0aca643-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.113780 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b5c68163-e0cf-43e8-8659-260fb0aca643-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.113915 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gpmf\" (UniqueName: \"kubernetes.io/projected/b5c68163-e0cf-43e8-8659-260fb0aca643-kube-api-access-5gpmf\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.114024 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b5c68163-e0cf-43e8-8659-260fb0aca643-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.114170 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.114288 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b5c68163-e0cf-43e8-8659-260fb0aca643-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.114436 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b5c68163-e0cf-43e8-8659-260fb0aca643-config-data\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.114562 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b5c68163-e0cf-43e8-8659-260fb0aca643-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.116913 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b5c68163-e0cf-43e8-8659-260fb0aca643-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.117737 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b5c68163-e0cf-43e8-8659-260fb0aca643-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.118287 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b5c68163-e0cf-43e8-8659-260fb0aca643-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.118397 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b5c68163-e0cf-43e8-8659-260fb0aca643-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.118741 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b5c68163-e0cf-43e8-8659-260fb0aca643-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.118885 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b5c68163-e0cf-43e8-8659-260fb0aca643-config-data\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.120340 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b5c68163-e0cf-43e8-8659-260fb0aca643-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.121327 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b5c68163-e0cf-43e8-8659-260fb0aca643-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.121488 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.135397 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b5c68163-e0cf-43e8-8659-260fb0aca643-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.139064 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gpmf\" (UniqueName: \"kubernetes.io/projected/b5c68163-e0cf-43e8-8659-260fb0aca643-kube-api-access-5gpmf\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.169305 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"b5c68163-e0cf-43e8-8659-260fb0aca643\") " pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.261697 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.519801 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.623661 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k92rn\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-kube-api-access-k92rn\") pod \"75dbebbf-c4f3-4508-901d-05148fd48f74\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.623737 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75dbebbf-c4f3-4508-901d-05148fd48f74-pod-info\") pod \"75dbebbf-c4f3-4508-901d-05148fd48f74\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.623918 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-plugins\") pod \"75dbebbf-c4f3-4508-901d-05148fd48f74\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.624351 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "75dbebbf-c4f3-4508-901d-05148fd48f74" (UID: "75dbebbf-c4f3-4508-901d-05148fd48f74"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.624506 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-server-conf\") pod \"75dbebbf-c4f3-4508-901d-05148fd48f74\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.624823 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75dbebbf-c4f3-4508-901d-05148fd48f74-erlang-cookie-secret\") pod \"75dbebbf-c4f3-4508-901d-05148fd48f74\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.625268 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"75dbebbf-c4f3-4508-901d-05148fd48f74\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.625541 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-tls\") pod \"75dbebbf-c4f3-4508-901d-05148fd48f74\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.625793 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-confd\") pod \"75dbebbf-c4f3-4508-901d-05148fd48f74\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.625822 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-plugins-conf\") pod \"75dbebbf-c4f3-4508-901d-05148fd48f74\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.625924 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-erlang-cookie\") pod \"75dbebbf-c4f3-4508-901d-05148fd48f74\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.625966 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-config-data\") pod \"75dbebbf-c4f3-4508-901d-05148fd48f74\" (UID: \"75dbebbf-c4f3-4508-901d-05148fd48f74\") " Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.626743 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.628667 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "75dbebbf-c4f3-4508-901d-05148fd48f74" (UID: "75dbebbf-c4f3-4508-901d-05148fd48f74"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.628745 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75dbebbf-c4f3-4508-901d-05148fd48f74-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "75dbebbf-c4f3-4508-901d-05148fd48f74" (UID: "75dbebbf-c4f3-4508-901d-05148fd48f74"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.629101 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "75dbebbf-c4f3-4508-901d-05148fd48f74" (UID: "75dbebbf-c4f3-4508-901d-05148fd48f74"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.630053 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "75dbebbf-c4f3-4508-901d-05148fd48f74" (UID: "75dbebbf-c4f3-4508-901d-05148fd48f74"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.630870 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-kube-api-access-k92rn" (OuterVolumeSpecName: "kube-api-access-k92rn") pod "75dbebbf-c4f3-4508-901d-05148fd48f74" (UID: "75dbebbf-c4f3-4508-901d-05148fd48f74"). InnerVolumeSpecName "kube-api-access-k92rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.631329 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/75dbebbf-c4f3-4508-901d-05148fd48f74-pod-info" (OuterVolumeSpecName: "pod-info") pod "75dbebbf-c4f3-4508-901d-05148fd48f74" (UID: "75dbebbf-c4f3-4508-901d-05148fd48f74"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.631648 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "75dbebbf-c4f3-4508-901d-05148fd48f74" (UID: "75dbebbf-c4f3-4508-901d-05148fd48f74"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.654527 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-config-data" (OuterVolumeSpecName: "config-data") pod "75dbebbf-c4f3-4508-901d-05148fd48f74" (UID: "75dbebbf-c4f3-4508-901d-05148fd48f74"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.693067 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-server-conf" (OuterVolumeSpecName: "server-conf") pod "75dbebbf-c4f3-4508-901d-05148fd48f74" (UID: "75dbebbf-c4f3-4508-901d-05148fd48f74"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.728421 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.728458 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k92rn\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-kube-api-access-k92rn\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.728472 4726 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75dbebbf-c4f3-4508-901d-05148fd48f74-pod-info\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.728485 4726 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-server-conf\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.728501 4726 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75dbebbf-c4f3-4508-901d-05148fd48f74-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.728534 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.728545 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.728556 4726 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75dbebbf-c4f3-4508-901d-05148fd48f74-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.728569 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.751580 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.758932 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "75dbebbf-c4f3-4508-901d-05148fd48f74" (UID: "75dbebbf-c4f3-4508-901d-05148fd48f74"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.815155 4726 generic.go:334] "Generic (PLEG): container finished" podID="75dbebbf-c4f3-4508-901d-05148fd48f74" containerID="cb8f6b26ed1180f41f0273e2461ce4276f12488072d6c566ce51a960f48e6ec2" exitCode=0 Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.815201 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.815218 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75dbebbf-c4f3-4508-901d-05148fd48f74","Type":"ContainerDied","Data":"cb8f6b26ed1180f41f0273e2461ce4276f12488072d6c566ce51a960f48e6ec2"} Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.816269 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75dbebbf-c4f3-4508-901d-05148fd48f74","Type":"ContainerDied","Data":"787e96333c99acae7d2829130a3c13c0c3569051acd1fa48211b55442a8b83c8"} Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.816293 4726 scope.go:117] "RemoveContainer" containerID="cb8f6b26ed1180f41f0273e2461ce4276f12488072d6c566ce51a960f48e6ec2" Oct 01 16:21:16 crc kubenswrapper[4726]: W1001 16:21:16.821462 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5c68163_e0cf_43e8_8659_260fb0aca643.slice/crio-79aa13dc157ae358494f1c5b3afec0cfb44234de43d518bde244c0045d970d4f WatchSource:0}: Error finding container 79aa13dc157ae358494f1c5b3afec0cfb44234de43d518bde244c0045d970d4f: Status 404 returned error can't find the container with id 79aa13dc157ae358494f1c5b3afec0cfb44234de43d518bde244c0045d970d4f Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.822247 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.829426 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.829551 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75dbebbf-c4f3-4508-901d-05148fd48f74-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.838623 4726 scope.go:117] "RemoveContainer" containerID="198fb4a2291752198257d14ae7da6349daa4258c3f409a6a3f85c2d745acc28d" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.859777 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.909679 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.935615 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:21:16 crc kubenswrapper[4726]: E1001 16:21:16.936115 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75dbebbf-c4f3-4508-901d-05148fd48f74" containerName="setup-container" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.936134 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="75dbebbf-c4f3-4508-901d-05148fd48f74" containerName="setup-container" Oct 01 16:21:16 crc kubenswrapper[4726]: E1001 16:21:16.936159 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75dbebbf-c4f3-4508-901d-05148fd48f74" containerName="rabbitmq" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.936168 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="75dbebbf-c4f3-4508-901d-05148fd48f74" containerName="rabbitmq" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.936409 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="75dbebbf-c4f3-4508-901d-05148fd48f74" containerName="rabbitmq" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.937821 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.942336 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.942384 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.942687 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.942791 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.942853 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.942953 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.943909 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-g5k8t" Oct 01 16:21:16 crc kubenswrapper[4726]: I1001 16:21:16.946256 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.007351 4726 scope.go:117] "RemoveContainer" containerID="cb8f6b26ed1180f41f0273e2461ce4276f12488072d6c566ce51a960f48e6ec2" Oct 01 16:21:17 crc kubenswrapper[4726]: E1001 16:21:17.007998 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb8f6b26ed1180f41f0273e2461ce4276f12488072d6c566ce51a960f48e6ec2\": container with ID starting with cb8f6b26ed1180f41f0273e2461ce4276f12488072d6c566ce51a960f48e6ec2 not found: ID does not exist" containerID="cb8f6b26ed1180f41f0273e2461ce4276f12488072d6c566ce51a960f48e6ec2" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.008060 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb8f6b26ed1180f41f0273e2461ce4276f12488072d6c566ce51a960f48e6ec2"} err="failed to get container status \"cb8f6b26ed1180f41f0273e2461ce4276f12488072d6c566ce51a960f48e6ec2\": rpc error: code = NotFound desc = could not find container \"cb8f6b26ed1180f41f0273e2461ce4276f12488072d6c566ce51a960f48e6ec2\": container with ID starting with cb8f6b26ed1180f41f0273e2461ce4276f12488072d6c566ce51a960f48e6ec2 not found: ID does not exist" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.008185 4726 scope.go:117] "RemoveContainer" containerID="198fb4a2291752198257d14ae7da6349daa4258c3f409a6a3f85c2d745acc28d" Oct 01 16:21:17 crc kubenswrapper[4726]: E1001 16:21:17.009618 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"198fb4a2291752198257d14ae7da6349daa4258c3f409a6a3f85c2d745acc28d\": container with ID starting with 198fb4a2291752198257d14ae7da6349daa4258c3f409a6a3f85c2d745acc28d not found: ID does not exist" containerID="198fb4a2291752198257d14ae7da6349daa4258c3f409a6a3f85c2d745acc28d" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.009867 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"198fb4a2291752198257d14ae7da6349daa4258c3f409a6a3f85c2d745acc28d"} err="failed to get container status \"198fb4a2291752198257d14ae7da6349daa4258c3f409a6a3f85c2d745acc28d\": rpc error: code = NotFound desc = could not find container \"198fb4a2291752198257d14ae7da6349daa4258c3f409a6a3f85c2d745acc28d\": container with ID starting with 198fb4a2291752198257d14ae7da6349daa4258c3f409a6a3f85c2d745acc28d not found: ID does not exist" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.036802 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8885e9d5-6b72-496e-b395-d0f94b3b991c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.036887 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.036955 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8885e9d5-6b72-496e-b395-d0f94b3b991c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.037010 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9nwq\" (UniqueName: \"kubernetes.io/projected/8885e9d5-6b72-496e-b395-d0f94b3b991c-kube-api-access-k9nwq\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.037042 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8885e9d5-6b72-496e-b395-d0f94b3b991c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.037086 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8885e9d5-6b72-496e-b395-d0f94b3b991c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.037134 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8885e9d5-6b72-496e-b395-d0f94b3b991c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.037154 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8885e9d5-6b72-496e-b395-d0f94b3b991c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.037221 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8885e9d5-6b72-496e-b395-d0f94b3b991c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.037266 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8885e9d5-6b72-496e-b395-d0f94b3b991c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.037352 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8885e9d5-6b72-496e-b395-d0f94b3b991c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.138587 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8885e9d5-6b72-496e-b395-d0f94b3b991c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.138627 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8885e9d5-6b72-496e-b395-d0f94b3b991c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.138655 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8885e9d5-6b72-496e-b395-d0f94b3b991c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.138680 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8885e9d5-6b72-496e-b395-d0f94b3b991c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.138734 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8885e9d5-6b72-496e-b395-d0f94b3b991c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.138753 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8885e9d5-6b72-496e-b395-d0f94b3b991c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.138788 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.138818 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8885e9d5-6b72-496e-b395-d0f94b3b991c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.138853 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9nwq\" (UniqueName: \"kubernetes.io/projected/8885e9d5-6b72-496e-b395-d0f94b3b991c-kube-api-access-k9nwq\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.138875 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8885e9d5-6b72-496e-b395-d0f94b3b991c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.138891 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8885e9d5-6b72-496e-b395-d0f94b3b991c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.139103 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8885e9d5-6b72-496e-b395-d0f94b3b991c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.139746 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.140043 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8885e9d5-6b72-496e-b395-d0f94b3b991c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.140177 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8885e9d5-6b72-496e-b395-d0f94b3b991c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.140433 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8885e9d5-6b72-496e-b395-d0f94b3b991c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.140491 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8885e9d5-6b72-496e-b395-d0f94b3b991c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.144417 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8885e9d5-6b72-496e-b395-d0f94b3b991c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.144860 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8885e9d5-6b72-496e-b395-d0f94b3b991c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.145480 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8885e9d5-6b72-496e-b395-d0f94b3b991c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.151920 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8885e9d5-6b72-496e-b395-d0f94b3b991c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.157810 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9nwq\" (UniqueName: \"kubernetes.io/projected/8885e9d5-6b72-496e-b395-d0f94b3b991c-kube-api-access-k9nwq\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.175351 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8885e9d5-6b72-496e-b395-d0f94b3b991c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.365804 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.799710 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.848840 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57593043-8bb5-4967-8a6c-1a55a3e8894c" path="/var/lib/kubelet/pods/57593043-8bb5-4967-8a6c-1a55a3e8894c/volumes" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.849663 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75dbebbf-c4f3-4508-901d-05148fd48f74" path="/var/lib/kubelet/pods/75dbebbf-c4f3-4508-901d-05148fd48f74/volumes" Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.877744 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8885e9d5-6b72-496e-b395-d0f94b3b991c","Type":"ContainerStarted","Data":"f211381265f6c9c814f02ed993f2936199af0c90482e5addfa10a881f00bace1"} Oct 01 16:21:17 crc kubenswrapper[4726]: I1001 16:21:17.884366 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b5c68163-e0cf-43e8-8659-260fb0aca643","Type":"ContainerStarted","Data":"79aa13dc157ae358494f1c5b3afec0cfb44234de43d518bde244c0045d970d4f"} Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.522286 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-dvlpp"] Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.526002 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.534125 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.565090 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-dvlpp"] Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.666897 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-dns-svc\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.667024 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.667247 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-config\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.667571 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9bbm\" (UniqueName: \"kubernetes.io/projected/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-kube-api-access-x9bbm\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.667736 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.667797 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.668193 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.770458 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.770521 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-dns-svc\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.770560 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.770604 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-config\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.770668 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9bbm\" (UniqueName: \"kubernetes.io/projected/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-kube-api-access-x9bbm\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.770715 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.770737 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.771915 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.772733 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.773937 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-config\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.774525 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-dns-svc\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.774822 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.775089 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.791100 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9bbm\" (UniqueName: \"kubernetes.io/projected/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-kube-api-access-x9bbm\") pod \"dnsmasq-dns-d558885bc-dvlpp\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.860275 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:18 crc kubenswrapper[4726]: I1001 16:21:18.910932 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b5c68163-e0cf-43e8-8659-260fb0aca643","Type":"ContainerStarted","Data":"181d6e603e62e92f4a6957d7441743c425fe78073b7d4ffc2dbf95480a0b8901"} Oct 01 16:21:19 crc kubenswrapper[4726]: I1001 16:21:19.372986 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-dvlpp"] Oct 01 16:21:19 crc kubenswrapper[4726]: W1001 16:21:19.381680 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3ecfc15_088b_46de_aa8c_84a3dd3abd81.slice/crio-c599b6abc1cd34b2790b8fa990a47ecbbf28e8f54a80eddc1c2f2bd46857255b WatchSource:0}: Error finding container c599b6abc1cd34b2790b8fa990a47ecbbf28e8f54a80eddc1c2f2bd46857255b: Status 404 returned error can't find the container with id c599b6abc1cd34b2790b8fa990a47ecbbf28e8f54a80eddc1c2f2bd46857255b Oct 01 16:21:19 crc kubenswrapper[4726]: I1001 16:21:19.923848 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-dvlpp" event={"ID":"d3ecfc15-088b-46de-aa8c-84a3dd3abd81","Type":"ContainerStarted","Data":"41c541d53c8c4d917c6f0567b8aa64f9a3e7539a0f67372075be823a0c3d478f"} Oct 01 16:21:19 crc kubenswrapper[4726]: I1001 16:21:19.924204 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-dvlpp" event={"ID":"d3ecfc15-088b-46de-aa8c-84a3dd3abd81","Type":"ContainerStarted","Data":"c599b6abc1cd34b2790b8fa990a47ecbbf28e8f54a80eddc1c2f2bd46857255b"} Oct 01 16:21:19 crc kubenswrapper[4726]: I1001 16:21:19.925638 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8885e9d5-6b72-496e-b395-d0f94b3b991c","Type":"ContainerStarted","Data":"38bdae26d4ce4665373870c477a04b5dcb0c18ce9bcb18f8447f4cf4dacfbfb4"} Oct 01 16:21:20 crc kubenswrapper[4726]: I1001 16:21:20.942758 4726 generic.go:334] "Generic (PLEG): container finished" podID="d3ecfc15-088b-46de-aa8c-84a3dd3abd81" containerID="41c541d53c8c4d917c6f0567b8aa64f9a3e7539a0f67372075be823a0c3d478f" exitCode=0 Oct 01 16:21:20 crc kubenswrapper[4726]: I1001 16:21:20.946571 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-dvlpp" event={"ID":"d3ecfc15-088b-46de-aa8c-84a3dd3abd81","Type":"ContainerDied","Data":"41c541d53c8c4d917c6f0567b8aa64f9a3e7539a0f67372075be823a0c3d478f"} Oct 01 16:21:21 crc kubenswrapper[4726]: I1001 16:21:21.961487 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-dvlpp" event={"ID":"d3ecfc15-088b-46de-aa8c-84a3dd3abd81","Type":"ContainerStarted","Data":"7decefab8179681e841b7d9df4b40b733be3d384e598f3e05fd627d7e974db51"} Oct 01 16:21:21 crc kubenswrapper[4726]: I1001 16:21:21.961991 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:21 crc kubenswrapper[4726]: I1001 16:21:21.983620 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-dvlpp" podStartSLOduration=3.98360868 podStartE2EDuration="3.98360868s" podCreationTimestamp="2025-10-01 16:21:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:21:21.978337278 +0000 UTC m=+1334.879889865" watchObservedRunningTime="2025-10-01 16:21:21.98360868 +0000 UTC m=+1334.885161257" Oct 01 16:21:23 crc kubenswrapper[4726]: I1001 16:21:23.414196 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:21:23 crc kubenswrapper[4726]: I1001 16:21:23.414271 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:21:28 crc kubenswrapper[4726]: I1001 16:21:28.862454 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:28 crc kubenswrapper[4726]: I1001 16:21:28.926852 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-jwdxg"] Oct 01 16:21:28 crc kubenswrapper[4726]: I1001 16:21:28.927174 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" podUID="70a049b2-bfa9-4ccf-bddb-ece60978a6be" containerName="dnsmasq-dns" containerID="cri-o://f13efaa67ae6b7e10b323aa5a532f9cff48c69cf9d95c5cfded2a150b5cc4531" gracePeriod=10 Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.106851 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-cx7ps"] Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.108842 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.119966 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-cx7ps"] Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.196679 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.196751 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.196804 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.196886 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.196937 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-config\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.196973 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j6fm\" (UniqueName: \"kubernetes.io/projected/09379814-0816-4c10-ba61-5d2875e3841a-kube-api-access-6j6fm\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.197007 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.298844 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.298913 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.298946 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.298991 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.299017 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-config\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.299050 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j6fm\" (UniqueName: \"kubernetes.io/projected/09379814-0816-4c10-ba61-5d2875e3841a-kube-api-access-6j6fm\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.299105 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.300276 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.300319 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.300357 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.300905 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-config\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.301082 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.301681 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09379814-0816-4c10-ba61-5d2875e3841a-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.330556 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j6fm\" (UniqueName: \"kubernetes.io/projected/09379814-0816-4c10-ba61-5d2875e3841a-kube-api-access-6j6fm\") pod \"dnsmasq-dns-78c64bc9c5-cx7ps\" (UID: \"09379814-0816-4c10-ba61-5d2875e3841a\") " pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.433992 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.436204 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.503244 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-dns-svc\") pod \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.503602 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-ovsdbserver-sb\") pod \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.503640 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tl6d\" (UniqueName: \"kubernetes.io/projected/70a049b2-bfa9-4ccf-bddb-ece60978a6be-kube-api-access-5tl6d\") pod \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.503690 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-config\") pod \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.503738 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-ovsdbserver-nb\") pod \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.503754 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-dns-swift-storage-0\") pod \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\" (UID: \"70a049b2-bfa9-4ccf-bddb-ece60978a6be\") " Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.510011 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70a049b2-bfa9-4ccf-bddb-ece60978a6be-kube-api-access-5tl6d" (OuterVolumeSpecName: "kube-api-access-5tl6d") pod "70a049b2-bfa9-4ccf-bddb-ece60978a6be" (UID: "70a049b2-bfa9-4ccf-bddb-ece60978a6be"). InnerVolumeSpecName "kube-api-access-5tl6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.566808 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "70a049b2-bfa9-4ccf-bddb-ece60978a6be" (UID: "70a049b2-bfa9-4ccf-bddb-ece60978a6be"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.570809 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "70a049b2-bfa9-4ccf-bddb-ece60978a6be" (UID: "70a049b2-bfa9-4ccf-bddb-ece60978a6be"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.578344 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "70a049b2-bfa9-4ccf-bddb-ece60978a6be" (UID: "70a049b2-bfa9-4ccf-bddb-ece60978a6be"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.582500 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-config" (OuterVolumeSpecName: "config") pod "70a049b2-bfa9-4ccf-bddb-ece60978a6be" (UID: "70a049b2-bfa9-4ccf-bddb-ece60978a6be"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.583414 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "70a049b2-bfa9-4ccf-bddb-ece60978a6be" (UID: "70a049b2-bfa9-4ccf-bddb-ece60978a6be"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.609284 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.609307 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.609330 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tl6d\" (UniqueName: \"kubernetes.io/projected/70a049b2-bfa9-4ccf-bddb-ece60978a6be-kube-api-access-5tl6d\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.609339 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.609347 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.609357 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/70a049b2-bfa9-4ccf-bddb-ece60978a6be-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:29 crc kubenswrapper[4726]: I1001 16:21:29.916620 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-cx7ps"] Oct 01 16:21:30 crc kubenswrapper[4726]: I1001 16:21:30.047213 4726 generic.go:334] "Generic (PLEG): container finished" podID="70a049b2-bfa9-4ccf-bddb-ece60978a6be" containerID="f13efaa67ae6b7e10b323aa5a532f9cff48c69cf9d95c5cfded2a150b5cc4531" exitCode=0 Oct 01 16:21:30 crc kubenswrapper[4726]: I1001 16:21:30.047300 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" Oct 01 16:21:30 crc kubenswrapper[4726]: I1001 16:21:30.047322 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" event={"ID":"70a049b2-bfa9-4ccf-bddb-ece60978a6be","Type":"ContainerDied","Data":"f13efaa67ae6b7e10b323aa5a532f9cff48c69cf9d95c5cfded2a150b5cc4531"} Oct 01 16:21:30 crc kubenswrapper[4726]: I1001 16:21:30.048034 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-jwdxg" event={"ID":"70a049b2-bfa9-4ccf-bddb-ece60978a6be","Type":"ContainerDied","Data":"94fd35abc08c223b29a56c6d04645d91598aa032d8bc4f208b31b7ab227a6ef2"} Oct 01 16:21:30 crc kubenswrapper[4726]: I1001 16:21:30.048081 4726 scope.go:117] "RemoveContainer" containerID="f13efaa67ae6b7e10b323aa5a532f9cff48c69cf9d95c5cfded2a150b5cc4531" Oct 01 16:21:30 crc kubenswrapper[4726]: I1001 16:21:30.051454 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" event={"ID":"09379814-0816-4c10-ba61-5d2875e3841a","Type":"ContainerStarted","Data":"6ad01d87556fdc333d48be766d6b8e31b1a2c235d8fd6ba22d6881bb295ec92b"} Oct 01 16:21:30 crc kubenswrapper[4726]: I1001 16:21:30.151783 4726 scope.go:117] "RemoveContainer" containerID="e6dfbc8d83b3bb5bb8192b054e6ab08f74173fb11f367fdd07b7ca9d0e002008" Oct 01 16:21:30 crc kubenswrapper[4726]: I1001 16:21:30.155461 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-jwdxg"] Oct 01 16:21:30 crc kubenswrapper[4726]: I1001 16:21:30.164141 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-jwdxg"] Oct 01 16:21:30 crc kubenswrapper[4726]: I1001 16:21:30.175698 4726 scope.go:117] "RemoveContainer" containerID="f13efaa67ae6b7e10b323aa5a532f9cff48c69cf9d95c5cfded2a150b5cc4531" Oct 01 16:21:30 crc kubenswrapper[4726]: E1001 16:21:30.176244 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f13efaa67ae6b7e10b323aa5a532f9cff48c69cf9d95c5cfded2a150b5cc4531\": container with ID starting with f13efaa67ae6b7e10b323aa5a532f9cff48c69cf9d95c5cfded2a150b5cc4531 not found: ID does not exist" containerID="f13efaa67ae6b7e10b323aa5a532f9cff48c69cf9d95c5cfded2a150b5cc4531" Oct 01 16:21:30 crc kubenswrapper[4726]: I1001 16:21:30.176298 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f13efaa67ae6b7e10b323aa5a532f9cff48c69cf9d95c5cfded2a150b5cc4531"} err="failed to get container status \"f13efaa67ae6b7e10b323aa5a532f9cff48c69cf9d95c5cfded2a150b5cc4531\": rpc error: code = NotFound desc = could not find container \"f13efaa67ae6b7e10b323aa5a532f9cff48c69cf9d95c5cfded2a150b5cc4531\": container with ID starting with f13efaa67ae6b7e10b323aa5a532f9cff48c69cf9d95c5cfded2a150b5cc4531 not found: ID does not exist" Oct 01 16:21:30 crc kubenswrapper[4726]: I1001 16:21:30.176323 4726 scope.go:117] "RemoveContainer" containerID="e6dfbc8d83b3bb5bb8192b054e6ab08f74173fb11f367fdd07b7ca9d0e002008" Oct 01 16:21:30 crc kubenswrapper[4726]: E1001 16:21:30.176805 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6dfbc8d83b3bb5bb8192b054e6ab08f74173fb11f367fdd07b7ca9d0e002008\": container with ID starting with e6dfbc8d83b3bb5bb8192b054e6ab08f74173fb11f367fdd07b7ca9d0e002008 not found: ID does not exist" containerID="e6dfbc8d83b3bb5bb8192b054e6ab08f74173fb11f367fdd07b7ca9d0e002008" Oct 01 16:21:30 crc kubenswrapper[4726]: I1001 16:21:30.176836 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6dfbc8d83b3bb5bb8192b054e6ab08f74173fb11f367fdd07b7ca9d0e002008"} err="failed to get container status \"e6dfbc8d83b3bb5bb8192b054e6ab08f74173fb11f367fdd07b7ca9d0e002008\": rpc error: code = NotFound desc = could not find container \"e6dfbc8d83b3bb5bb8192b054e6ab08f74173fb11f367fdd07b7ca9d0e002008\": container with ID starting with e6dfbc8d83b3bb5bb8192b054e6ab08f74173fb11f367fdd07b7ca9d0e002008 not found: ID does not exist" Oct 01 16:21:31 crc kubenswrapper[4726]: I1001 16:21:31.063516 4726 generic.go:334] "Generic (PLEG): container finished" podID="09379814-0816-4c10-ba61-5d2875e3841a" containerID="7d1ab3845be66063e971c04c071ca9f5f08795bd571ad81b7c81696e03189cfc" exitCode=0 Oct 01 16:21:31 crc kubenswrapper[4726]: I1001 16:21:31.063623 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" event={"ID":"09379814-0816-4c10-ba61-5d2875e3841a","Type":"ContainerDied","Data":"7d1ab3845be66063e971c04c071ca9f5f08795bd571ad81b7c81696e03189cfc"} Oct 01 16:21:31 crc kubenswrapper[4726]: I1001 16:21:31.822276 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70a049b2-bfa9-4ccf-bddb-ece60978a6be" path="/var/lib/kubelet/pods/70a049b2-bfa9-4ccf-bddb-ece60978a6be/volumes" Oct 01 16:21:32 crc kubenswrapper[4726]: I1001 16:21:32.081508 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" event={"ID":"09379814-0816-4c10-ba61-5d2875e3841a","Type":"ContainerStarted","Data":"b6a66d1d1214798595d2c158da2d0eedb7f5f197f859a958fdb0bb4eb667e19a"} Oct 01 16:21:32 crc kubenswrapper[4726]: I1001 16:21:32.081664 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:32 crc kubenswrapper[4726]: I1001 16:21:32.097007 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" podStartSLOduration=3.096989255 podStartE2EDuration="3.096989255s" podCreationTimestamp="2025-10-01 16:21:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:21:32.096168921 +0000 UTC m=+1344.997721588" watchObservedRunningTime="2025-10-01 16:21:32.096989255 +0000 UTC m=+1344.998541832" Oct 01 16:21:39 crc kubenswrapper[4726]: I1001 16:21:39.435366 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78c64bc9c5-cx7ps" Oct 01 16:21:39 crc kubenswrapper[4726]: I1001 16:21:39.558660 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-dvlpp"] Oct 01 16:21:39 crc kubenswrapper[4726]: I1001 16:21:39.558930 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-dvlpp" podUID="d3ecfc15-088b-46de-aa8c-84a3dd3abd81" containerName="dnsmasq-dns" containerID="cri-o://7decefab8179681e841b7d9df4b40b733be3d384e598f3e05fd627d7e974db51" gracePeriod=10 Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.032934 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.133067 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-openstack-edpm-ipam\") pod \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.133160 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-ovsdbserver-nb\") pod \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.133336 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-config\") pod \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.133425 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9bbm\" (UniqueName: \"kubernetes.io/projected/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-kube-api-access-x9bbm\") pod \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.133667 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-dns-swift-storage-0\") pod \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.133722 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-dns-svc\") pod \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.133746 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-ovsdbserver-sb\") pod \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\" (UID: \"d3ecfc15-088b-46de-aa8c-84a3dd3abd81\") " Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.138133 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-kube-api-access-x9bbm" (OuterVolumeSpecName: "kube-api-access-x9bbm") pod "d3ecfc15-088b-46de-aa8c-84a3dd3abd81" (UID: "d3ecfc15-088b-46de-aa8c-84a3dd3abd81"). InnerVolumeSpecName "kube-api-access-x9bbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.172284 4726 generic.go:334] "Generic (PLEG): container finished" podID="d3ecfc15-088b-46de-aa8c-84a3dd3abd81" containerID="7decefab8179681e841b7d9df4b40b733be3d384e598f3e05fd627d7e974db51" exitCode=0 Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.172325 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-dvlpp" event={"ID":"d3ecfc15-088b-46de-aa8c-84a3dd3abd81","Type":"ContainerDied","Data":"7decefab8179681e841b7d9df4b40b733be3d384e598f3e05fd627d7e974db51"} Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.172351 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-dvlpp" event={"ID":"d3ecfc15-088b-46de-aa8c-84a3dd3abd81","Type":"ContainerDied","Data":"c599b6abc1cd34b2790b8fa990a47ecbbf28e8f54a80eddc1c2f2bd46857255b"} Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.172368 4726 scope.go:117] "RemoveContainer" containerID="7decefab8179681e841b7d9df4b40b733be3d384e598f3e05fd627d7e974db51" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.172691 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-dvlpp" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.185123 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d3ecfc15-088b-46de-aa8c-84a3dd3abd81" (UID: "d3ecfc15-088b-46de-aa8c-84a3dd3abd81"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.188317 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "d3ecfc15-088b-46de-aa8c-84a3dd3abd81" (UID: "d3ecfc15-088b-46de-aa8c-84a3dd3abd81"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.202630 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d3ecfc15-088b-46de-aa8c-84a3dd3abd81" (UID: "d3ecfc15-088b-46de-aa8c-84a3dd3abd81"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.203542 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d3ecfc15-088b-46de-aa8c-84a3dd3abd81" (UID: "d3ecfc15-088b-46de-aa8c-84a3dd3abd81"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.209355 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-config" (OuterVolumeSpecName: "config") pod "d3ecfc15-088b-46de-aa8c-84a3dd3abd81" (UID: "d3ecfc15-088b-46de-aa8c-84a3dd3abd81"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.210349 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d3ecfc15-088b-46de-aa8c-84a3dd3abd81" (UID: "d3ecfc15-088b-46de-aa8c-84a3dd3abd81"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.236301 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9bbm\" (UniqueName: \"kubernetes.io/projected/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-kube-api-access-x9bbm\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.236668 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.236756 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.236916 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.237113 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.237292 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.237879 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ecfc15-088b-46de-aa8c-84a3dd3abd81-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.278119 4726 scope.go:117] "RemoveContainer" containerID="41c541d53c8c4d917c6f0567b8aa64f9a3e7539a0f67372075be823a0c3d478f" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.296849 4726 scope.go:117] "RemoveContainer" containerID="7decefab8179681e841b7d9df4b40b733be3d384e598f3e05fd627d7e974db51" Oct 01 16:21:40 crc kubenswrapper[4726]: E1001 16:21:40.297429 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7decefab8179681e841b7d9df4b40b733be3d384e598f3e05fd627d7e974db51\": container with ID starting with 7decefab8179681e841b7d9df4b40b733be3d384e598f3e05fd627d7e974db51 not found: ID does not exist" containerID="7decefab8179681e841b7d9df4b40b733be3d384e598f3e05fd627d7e974db51" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.297539 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7decefab8179681e841b7d9df4b40b733be3d384e598f3e05fd627d7e974db51"} err="failed to get container status \"7decefab8179681e841b7d9df4b40b733be3d384e598f3e05fd627d7e974db51\": rpc error: code = NotFound desc = could not find container \"7decefab8179681e841b7d9df4b40b733be3d384e598f3e05fd627d7e974db51\": container with ID starting with 7decefab8179681e841b7d9df4b40b733be3d384e598f3e05fd627d7e974db51 not found: ID does not exist" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.297639 4726 scope.go:117] "RemoveContainer" containerID="41c541d53c8c4d917c6f0567b8aa64f9a3e7539a0f67372075be823a0c3d478f" Oct 01 16:21:40 crc kubenswrapper[4726]: E1001 16:21:40.297982 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41c541d53c8c4d917c6f0567b8aa64f9a3e7539a0f67372075be823a0c3d478f\": container with ID starting with 41c541d53c8c4d917c6f0567b8aa64f9a3e7539a0f67372075be823a0c3d478f not found: ID does not exist" containerID="41c541d53c8c4d917c6f0567b8aa64f9a3e7539a0f67372075be823a0c3d478f" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.298096 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41c541d53c8c4d917c6f0567b8aa64f9a3e7539a0f67372075be823a0c3d478f"} err="failed to get container status \"41c541d53c8c4d917c6f0567b8aa64f9a3e7539a0f67372075be823a0c3d478f\": rpc error: code = NotFound desc = could not find container \"41c541d53c8c4d917c6f0567b8aa64f9a3e7539a0f67372075be823a0c3d478f\": container with ID starting with 41c541d53c8c4d917c6f0567b8aa64f9a3e7539a0f67372075be823a0c3d478f not found: ID does not exist" Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.524642 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-dvlpp"] Oct 01 16:21:40 crc kubenswrapper[4726]: I1001 16:21:40.536520 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-dvlpp"] Oct 01 16:21:41 crc kubenswrapper[4726]: I1001 16:21:41.819662 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3ecfc15-088b-46de-aa8c-84a3dd3abd81" path="/var/lib/kubelet/pods/d3ecfc15-088b-46de-aa8c-84a3dd3abd81/volumes" Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.273675 4726 generic.go:334] "Generic (PLEG): container finished" podID="b5c68163-e0cf-43e8-8659-260fb0aca643" containerID="181d6e603e62e92f4a6957d7441743c425fe78073b7d4ffc2dbf95480a0b8901" exitCode=0 Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.273775 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b5c68163-e0cf-43e8-8659-260fb0aca643","Type":"ContainerDied","Data":"181d6e603e62e92f4a6957d7441743c425fe78073b7d4ffc2dbf95480a0b8901"} Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.915338 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn"] Oct 01 16:21:51 crc kubenswrapper[4726]: E1001 16:21:51.916200 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a049b2-bfa9-4ccf-bddb-ece60978a6be" containerName="init" Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.916224 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a049b2-bfa9-4ccf-bddb-ece60978a6be" containerName="init" Oct 01 16:21:51 crc kubenswrapper[4726]: E1001 16:21:51.916244 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3ecfc15-088b-46de-aa8c-84a3dd3abd81" containerName="dnsmasq-dns" Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.916254 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3ecfc15-088b-46de-aa8c-84a3dd3abd81" containerName="dnsmasq-dns" Oct 01 16:21:51 crc kubenswrapper[4726]: E1001 16:21:51.916280 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3ecfc15-088b-46de-aa8c-84a3dd3abd81" containerName="init" Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.916288 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3ecfc15-088b-46de-aa8c-84a3dd3abd81" containerName="init" Oct 01 16:21:51 crc kubenswrapper[4726]: E1001 16:21:51.916310 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a049b2-bfa9-4ccf-bddb-ece60978a6be" containerName="dnsmasq-dns" Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.916319 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a049b2-bfa9-4ccf-bddb-ece60978a6be" containerName="dnsmasq-dns" Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.916538 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3ecfc15-088b-46de-aa8c-84a3dd3abd81" containerName="dnsmasq-dns" Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.916572 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="70a049b2-bfa9-4ccf-bddb-ece60978a6be" containerName="dnsmasq-dns" Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.928507 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.931234 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.931447 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.937536 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.937778 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:21:51 crc kubenswrapper[4726]: I1001 16:21:51.938943 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn"] Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.065568 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.065661 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjgzg\" (UniqueName: \"kubernetes.io/projected/01436a04-eda0-42cf-a799-29d2bddd197e-kube-api-access-hjgzg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.065774 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.065849 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.168176 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.168574 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjgzg\" (UniqueName: \"kubernetes.io/projected/01436a04-eda0-42cf-a799-29d2bddd197e-kube-api-access-hjgzg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.168672 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.168771 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.175160 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.175176 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.175859 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.189992 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjgzg\" (UniqueName: \"kubernetes.io/projected/01436a04-eda0-42cf-a799-29d2bddd197e-kube-api-access-hjgzg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.254144 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.285473 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b5c68163-e0cf-43e8-8659-260fb0aca643","Type":"ContainerStarted","Data":"783452947c7c84a73b100c6f7a305437315760d029dfc837f55a194a16c3d528"} Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.286536 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.840654 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.840618548 podStartE2EDuration="37.840618548s" podCreationTimestamp="2025-10-01 16:21:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:21:52.323736535 +0000 UTC m=+1365.225289122" watchObservedRunningTime="2025-10-01 16:21:52.840618548 +0000 UTC m=+1365.742171155" Oct 01 16:21:52 crc kubenswrapper[4726]: I1001 16:21:52.849788 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn"] Oct 01 16:21:52 crc kubenswrapper[4726]: W1001 16:21:52.855541 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01436a04_eda0_42cf_a799_29d2bddd197e.slice/crio-c973c8d3c5c08fe27f1f990e3a3421a15bf7d38192c3f41deb1b65991be553fe WatchSource:0}: Error finding container c973c8d3c5c08fe27f1f990e3a3421a15bf7d38192c3f41deb1b65991be553fe: Status 404 returned error can't find the container with id c973c8d3c5c08fe27f1f990e3a3421a15bf7d38192c3f41deb1b65991be553fe Oct 01 16:21:53 crc kubenswrapper[4726]: I1001 16:21:53.295269 4726 generic.go:334] "Generic (PLEG): container finished" podID="8885e9d5-6b72-496e-b395-d0f94b3b991c" containerID="38bdae26d4ce4665373870c477a04b5dcb0c18ce9bcb18f8447f4cf4dacfbfb4" exitCode=0 Oct 01 16:21:53 crc kubenswrapper[4726]: I1001 16:21:53.295381 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8885e9d5-6b72-496e-b395-d0f94b3b991c","Type":"ContainerDied","Data":"38bdae26d4ce4665373870c477a04b5dcb0c18ce9bcb18f8447f4cf4dacfbfb4"} Oct 01 16:21:53 crc kubenswrapper[4726]: I1001 16:21:53.297348 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" event={"ID":"01436a04-eda0-42cf-a799-29d2bddd197e","Type":"ContainerStarted","Data":"c973c8d3c5c08fe27f1f990e3a3421a15bf7d38192c3f41deb1b65991be553fe"} Oct 01 16:21:53 crc kubenswrapper[4726]: I1001 16:21:53.415196 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:21:53 crc kubenswrapper[4726]: I1001 16:21:53.415453 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:21:53 crc kubenswrapper[4726]: I1001 16:21:53.415493 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 16:21:53 crc kubenswrapper[4726]: I1001 16:21:53.416141 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f2fff8833153004baace8bca4e9e33e62ef29d5bce54ddbea47fab14e44b1073"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:21:53 crc kubenswrapper[4726]: I1001 16:21:53.416190 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://f2fff8833153004baace8bca4e9e33e62ef29d5bce54ddbea47fab14e44b1073" gracePeriod=600 Oct 01 16:21:54 crc kubenswrapper[4726]: I1001 16:21:54.313893 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="f2fff8833153004baace8bca4e9e33e62ef29d5bce54ddbea47fab14e44b1073" exitCode=0 Oct 01 16:21:54 crc kubenswrapper[4726]: I1001 16:21:54.314031 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"f2fff8833153004baace8bca4e9e33e62ef29d5bce54ddbea47fab14e44b1073"} Oct 01 16:21:54 crc kubenswrapper[4726]: I1001 16:21:54.314270 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840"} Oct 01 16:21:54 crc kubenswrapper[4726]: I1001 16:21:54.314293 4726 scope.go:117] "RemoveContainer" containerID="db368996f7ef78489a378108debc56732f3e0a06eb79040d4b7e667cfd8503a8" Oct 01 16:21:54 crc kubenswrapper[4726]: I1001 16:21:54.322423 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8885e9d5-6b72-496e-b395-d0f94b3b991c","Type":"ContainerStarted","Data":"69e1bb1a875854f9a32724544392a4a7cc32739af632ef3609732ff7575edae9"} Oct 01 16:21:54 crc kubenswrapper[4726]: I1001 16:21:54.322802 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:21:54 crc kubenswrapper[4726]: I1001 16:21:54.361857 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.361841656 podStartE2EDuration="38.361841656s" podCreationTimestamp="2025-10-01 16:21:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:21:54.358986514 +0000 UTC m=+1367.260539091" watchObservedRunningTime="2025-10-01 16:21:54.361841656 +0000 UTC m=+1367.263394233" Oct 01 16:22:04 crc kubenswrapper[4726]: I1001 16:22:04.431853 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" event={"ID":"01436a04-eda0-42cf-a799-29d2bddd197e","Type":"ContainerStarted","Data":"b96d467048e17dc687bf5124d19206709bb99322dca2acfd32e1fe980579f8d8"} Oct 01 16:22:04 crc kubenswrapper[4726]: I1001 16:22:04.455632 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" podStartSLOduration=3.038834298 podStartE2EDuration="13.455614636s" podCreationTimestamp="2025-10-01 16:21:51 +0000 UTC" firstStartedPulling="2025-10-01 16:21:52.859480552 +0000 UTC m=+1365.761033129" lastFinishedPulling="2025-10-01 16:22:03.27626089 +0000 UTC m=+1376.177813467" observedRunningTime="2025-10-01 16:22:04.451164697 +0000 UTC m=+1377.352717284" watchObservedRunningTime="2025-10-01 16:22:04.455614636 +0000 UTC m=+1377.357167213" Oct 01 16:22:06 crc kubenswrapper[4726]: I1001 16:22:06.266254 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 01 16:22:07 crc kubenswrapper[4726]: I1001 16:22:07.369822 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:22:16 crc kubenswrapper[4726]: I1001 16:22:16.554015 4726 generic.go:334] "Generic (PLEG): container finished" podID="01436a04-eda0-42cf-a799-29d2bddd197e" containerID="b96d467048e17dc687bf5124d19206709bb99322dca2acfd32e1fe980579f8d8" exitCode=0 Oct 01 16:22:16 crc kubenswrapper[4726]: I1001 16:22:16.554144 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" event={"ID":"01436a04-eda0-42cf-a799-29d2bddd197e","Type":"ContainerDied","Data":"b96d467048e17dc687bf5124d19206709bb99322dca2acfd32e1fe980579f8d8"} Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.073117 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.115769 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-repo-setup-combined-ca-bundle\") pod \"01436a04-eda0-42cf-a799-29d2bddd197e\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.115851 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-inventory\") pod \"01436a04-eda0-42cf-a799-29d2bddd197e\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.116013 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjgzg\" (UniqueName: \"kubernetes.io/projected/01436a04-eda0-42cf-a799-29d2bddd197e-kube-api-access-hjgzg\") pod \"01436a04-eda0-42cf-a799-29d2bddd197e\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.116094 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-ssh-key\") pod \"01436a04-eda0-42cf-a799-29d2bddd197e\" (UID: \"01436a04-eda0-42cf-a799-29d2bddd197e\") " Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.127254 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "01436a04-eda0-42cf-a799-29d2bddd197e" (UID: "01436a04-eda0-42cf-a799-29d2bddd197e"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.127322 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01436a04-eda0-42cf-a799-29d2bddd197e-kube-api-access-hjgzg" (OuterVolumeSpecName: "kube-api-access-hjgzg") pod "01436a04-eda0-42cf-a799-29d2bddd197e" (UID: "01436a04-eda0-42cf-a799-29d2bddd197e"). InnerVolumeSpecName "kube-api-access-hjgzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.147974 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-inventory" (OuterVolumeSpecName: "inventory") pod "01436a04-eda0-42cf-a799-29d2bddd197e" (UID: "01436a04-eda0-42cf-a799-29d2bddd197e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.150918 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "01436a04-eda0-42cf-a799-29d2bddd197e" (UID: "01436a04-eda0-42cf-a799-29d2bddd197e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.218014 4726 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.218052 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.218063 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjgzg\" (UniqueName: \"kubernetes.io/projected/01436a04-eda0-42cf-a799-29d2bddd197e-kube-api-access-hjgzg\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.218071 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01436a04-eda0-42cf-a799-29d2bddd197e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.569974 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" event={"ID":"01436a04-eda0-42cf-a799-29d2bddd197e","Type":"ContainerDied","Data":"c973c8d3c5c08fe27f1f990e3a3421a15bf7d38192c3f41deb1b65991be553fe"} Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.570018 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c973c8d3c5c08fe27f1f990e3a3421a15bf7d38192c3f41deb1b65991be553fe" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.570103 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.651379 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g"] Oct 01 16:22:18 crc kubenswrapper[4726]: E1001 16:22:18.651807 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01436a04-eda0-42cf-a799-29d2bddd197e" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.651828 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="01436a04-eda0-42cf-a799-29d2bddd197e" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.652132 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="01436a04-eda0-42cf-a799-29d2bddd197e" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.652835 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.657156 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.657391 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.657576 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.657727 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.670450 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g"] Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.725979 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rsl2g\" (UID: \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.726030 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rsl2g\" (UID: \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.726295 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksf27\" (UniqueName: \"kubernetes.io/projected/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-kube-api-access-ksf27\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rsl2g\" (UID: \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.827859 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rsl2g\" (UID: \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.827933 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksf27\" (UniqueName: \"kubernetes.io/projected/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-kube-api-access-ksf27\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rsl2g\" (UID: \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.828252 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rsl2g\" (UID: \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.832061 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rsl2g\" (UID: \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.832959 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rsl2g\" (UID: \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.848690 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksf27\" (UniqueName: \"kubernetes.io/projected/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-kube-api-access-ksf27\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rsl2g\" (UID: \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" Oct 01 16:22:18 crc kubenswrapper[4726]: I1001 16:22:18.980862 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" Oct 01 16:22:19 crc kubenswrapper[4726]: I1001 16:22:19.514020 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g"] Oct 01 16:22:19 crc kubenswrapper[4726]: I1001 16:22:19.579332 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" event={"ID":"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3","Type":"ContainerStarted","Data":"62987be14359c59b6e43a86f10f8b5f1cab3f0efbfb3d76936d149195691c305"} Oct 01 16:22:20 crc kubenswrapper[4726]: I1001 16:22:20.590556 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" event={"ID":"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3","Type":"ContainerStarted","Data":"7c0330950ddb351108d7a36126a677ef1b11f0f983b0cf690fc7c8990d22d893"} Oct 01 16:22:20 crc kubenswrapper[4726]: I1001 16:22:20.613259 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" podStartSLOduration=2.031303006 podStartE2EDuration="2.613234436s" podCreationTimestamp="2025-10-01 16:22:18 +0000 UTC" firstStartedPulling="2025-10-01 16:22:19.519466219 +0000 UTC m=+1392.421018796" lastFinishedPulling="2025-10-01 16:22:20.101397649 +0000 UTC m=+1393.002950226" observedRunningTime="2025-10-01 16:22:20.604774262 +0000 UTC m=+1393.506326839" watchObservedRunningTime="2025-10-01 16:22:20.613234436 +0000 UTC m=+1393.514787013" Oct 01 16:22:21 crc kubenswrapper[4726]: I1001 16:22:21.240832 4726 scope.go:117] "RemoveContainer" containerID="2f52687b6d1be2679e117f20e531f467956a369117c02dd253f419d21137f23b" Oct 01 16:22:23 crc kubenswrapper[4726]: I1001 16:22:23.624683 4726 generic.go:334] "Generic (PLEG): container finished" podID="fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3" containerID="7c0330950ddb351108d7a36126a677ef1b11f0f983b0cf690fc7c8990d22d893" exitCode=0 Oct 01 16:22:23 crc kubenswrapper[4726]: I1001 16:22:23.624730 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" event={"ID":"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3","Type":"ContainerDied","Data":"7c0330950ddb351108d7a36126a677ef1b11f0f983b0cf690fc7c8990d22d893"} Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.093697 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.253364 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksf27\" (UniqueName: \"kubernetes.io/projected/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-kube-api-access-ksf27\") pod \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\" (UID: \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\") " Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.253429 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-inventory\") pod \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\" (UID: \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\") " Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.253509 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-ssh-key\") pod \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\" (UID: \"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3\") " Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.264666 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-kube-api-access-ksf27" (OuterVolumeSpecName: "kube-api-access-ksf27") pod "fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3" (UID: "fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3"). InnerVolumeSpecName "kube-api-access-ksf27". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.283927 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-inventory" (OuterVolumeSpecName: "inventory") pod "fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3" (UID: "fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.284329 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3" (UID: "fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.355168 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.355435 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksf27\" (UniqueName: \"kubernetes.io/projected/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-kube-api-access-ksf27\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.355448 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.656547 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" event={"ID":"fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3","Type":"ContainerDied","Data":"62987be14359c59b6e43a86f10f8b5f1cab3f0efbfb3d76936d149195691c305"} Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.656593 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62987be14359c59b6e43a86f10f8b5f1cab3f0efbfb3d76936d149195691c305" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.656618 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rsl2g" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.725164 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd"] Oct 01 16:22:25 crc kubenswrapper[4726]: E1001 16:22:25.725613 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.725632 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.725847 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.726525 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.729802 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.730037 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.730219 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.730522 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.736800 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd"] Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.763595 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.764015 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwfnm\" (UniqueName: \"kubernetes.io/projected/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-kube-api-access-jwfnm\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.764148 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.764205 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.866423 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwfnm\" (UniqueName: \"kubernetes.io/projected/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-kube-api-access-jwfnm\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.866471 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.866510 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.866562 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.872605 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.878768 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.880283 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:22:25 crc kubenswrapper[4726]: I1001 16:22:25.883539 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwfnm\" (UniqueName: \"kubernetes.io/projected/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-kube-api-access-jwfnm\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:22:26 crc kubenswrapper[4726]: I1001 16:22:26.050028 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:22:26 crc kubenswrapper[4726]: I1001 16:22:26.583924 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd"] Oct 01 16:22:26 crc kubenswrapper[4726]: W1001 16:22:26.585500 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf9f4ecc_09d0_482c_81d2_d7fe9503de1f.slice/crio-c2e519cfc09459945a82c96c3f1753582dd290313e93e7bba57614344f38c469 WatchSource:0}: Error finding container c2e519cfc09459945a82c96c3f1753582dd290313e93e7bba57614344f38c469: Status 404 returned error can't find the container with id c2e519cfc09459945a82c96c3f1753582dd290313e93e7bba57614344f38c469 Oct 01 16:22:26 crc kubenswrapper[4726]: I1001 16:22:26.666818 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" event={"ID":"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f","Type":"ContainerStarted","Data":"c2e519cfc09459945a82c96c3f1753582dd290313e93e7bba57614344f38c469"} Oct 01 16:22:36 crc kubenswrapper[4726]: I1001 16:22:36.813812 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" event={"ID":"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f","Type":"ContainerStarted","Data":"1d4bce1a8d60503608b9e886c2ef531096c709b261cec5f703135bef8185395a"} Oct 01 16:22:45 crc kubenswrapper[4726]: I1001 16:22:45.596479 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-57c9dff847-vbzrr" podUID="85ac310a-31b1-40c9-9724-2c488c0061e2" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 01 16:22:57 crc kubenswrapper[4726]: I1001 16:22:57.361361 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" podStartSLOduration=22.860917621 podStartE2EDuration="32.361332772s" podCreationTimestamp="2025-10-01 16:22:25 +0000 UTC" firstStartedPulling="2025-10-01 16:22:26.587472001 +0000 UTC m=+1399.489024578" lastFinishedPulling="2025-10-01 16:22:36.087887132 +0000 UTC m=+1408.989439729" observedRunningTime="2025-10-01 16:22:36.832583486 +0000 UTC m=+1409.734136123" watchObservedRunningTime="2025-10-01 16:22:57.361332772 +0000 UTC m=+1430.262885389" Oct 01 16:22:57 crc kubenswrapper[4726]: I1001 16:22:57.362969 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jjmn4"] Oct 01 16:22:57 crc kubenswrapper[4726]: I1001 16:22:57.367109 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:22:57 crc kubenswrapper[4726]: I1001 16:22:57.396798 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jjmn4"] Oct 01 16:22:57 crc kubenswrapper[4726]: I1001 16:22:57.502835 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm95l\" (UniqueName: \"kubernetes.io/projected/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-kube-api-access-dm95l\") pod \"redhat-marketplace-jjmn4\" (UID: \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\") " pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:22:57 crc kubenswrapper[4726]: I1001 16:22:57.502916 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-utilities\") pod \"redhat-marketplace-jjmn4\" (UID: \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\") " pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:22:57 crc kubenswrapper[4726]: I1001 16:22:57.502965 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-catalog-content\") pod \"redhat-marketplace-jjmn4\" (UID: \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\") " pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:22:57 crc kubenswrapper[4726]: I1001 16:22:57.604502 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm95l\" (UniqueName: \"kubernetes.io/projected/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-kube-api-access-dm95l\") pod \"redhat-marketplace-jjmn4\" (UID: \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\") " pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:22:57 crc kubenswrapper[4726]: I1001 16:22:57.604560 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-utilities\") pod \"redhat-marketplace-jjmn4\" (UID: \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\") " pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:22:57 crc kubenswrapper[4726]: I1001 16:22:57.604599 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-catalog-content\") pod \"redhat-marketplace-jjmn4\" (UID: \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\") " pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:22:57 crc kubenswrapper[4726]: I1001 16:22:57.605165 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-catalog-content\") pod \"redhat-marketplace-jjmn4\" (UID: \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\") " pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:22:57 crc kubenswrapper[4726]: I1001 16:22:57.605403 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-utilities\") pod \"redhat-marketplace-jjmn4\" (UID: \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\") " pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:22:57 crc kubenswrapper[4726]: I1001 16:22:57.625890 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm95l\" (UniqueName: \"kubernetes.io/projected/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-kube-api-access-dm95l\") pod \"redhat-marketplace-jjmn4\" (UID: \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\") " pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:22:57 crc kubenswrapper[4726]: I1001 16:22:57.694487 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:22:58 crc kubenswrapper[4726]: I1001 16:22:58.160646 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jjmn4"] Oct 01 16:22:59 crc kubenswrapper[4726]: I1001 16:22:59.037659 4726 generic.go:334] "Generic (PLEG): container finished" podID="06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" containerID="85281e5a919e1685d3fee888b4da0426daa6d668cdbd64b12a2660c2f924e7c2" exitCode=0 Oct 01 16:22:59 crc kubenswrapper[4726]: I1001 16:22:59.037772 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jjmn4" event={"ID":"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a","Type":"ContainerDied","Data":"85281e5a919e1685d3fee888b4da0426daa6d668cdbd64b12a2660c2f924e7c2"} Oct 01 16:22:59 crc kubenswrapper[4726]: I1001 16:22:59.037949 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jjmn4" event={"ID":"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a","Type":"ContainerStarted","Data":"183bde35048fa7aafe7a91e86ef32a3c9fcda182e78398841b5874e485adb8b7"} Oct 01 16:23:01 crc kubenswrapper[4726]: I1001 16:23:01.056733 4726 generic.go:334] "Generic (PLEG): container finished" podID="06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" containerID="2eebe96a9fd303d3b364239f425d4a95158561869f1049eb1776d932ea9185ec" exitCode=0 Oct 01 16:23:01 crc kubenswrapper[4726]: I1001 16:23:01.056868 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jjmn4" event={"ID":"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a","Type":"ContainerDied","Data":"2eebe96a9fd303d3b364239f425d4a95158561869f1049eb1776d932ea9185ec"} Oct 01 16:23:02 crc kubenswrapper[4726]: I1001 16:23:02.069158 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jjmn4" event={"ID":"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a","Type":"ContainerStarted","Data":"e76ea193ba740035f09291335dc73d7cdb81e7d2a57d6d22f1b284a97bb05d63"} Oct 01 16:23:02 crc kubenswrapper[4726]: I1001 16:23:02.094225 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jjmn4" podStartSLOduration=2.659186564 podStartE2EDuration="5.094207639s" podCreationTimestamp="2025-10-01 16:22:57 +0000 UTC" firstStartedPulling="2025-10-01 16:22:59.0415451 +0000 UTC m=+1431.943097677" lastFinishedPulling="2025-10-01 16:23:01.476566175 +0000 UTC m=+1434.378118752" observedRunningTime="2025-10-01 16:23:02.087362435 +0000 UTC m=+1434.988915032" watchObservedRunningTime="2025-10-01 16:23:02.094207639 +0000 UTC m=+1434.995760216" Oct 01 16:23:07 crc kubenswrapper[4726]: I1001 16:23:07.694660 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:23:07 crc kubenswrapper[4726]: I1001 16:23:07.695383 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:23:07 crc kubenswrapper[4726]: I1001 16:23:07.751139 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:23:08 crc kubenswrapper[4726]: I1001 16:23:08.192075 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:23:08 crc kubenswrapper[4726]: I1001 16:23:08.257159 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jjmn4"] Oct 01 16:23:10 crc kubenswrapper[4726]: I1001 16:23:10.158205 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jjmn4" podUID="06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" containerName="registry-server" containerID="cri-o://e76ea193ba740035f09291335dc73d7cdb81e7d2a57d6d22f1b284a97bb05d63" gracePeriod=2 Oct 01 16:23:10 crc kubenswrapper[4726]: I1001 16:23:10.685214 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:23:10 crc kubenswrapper[4726]: I1001 16:23:10.764580 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dm95l\" (UniqueName: \"kubernetes.io/projected/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-kube-api-access-dm95l\") pod \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\" (UID: \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\") " Oct 01 16:23:10 crc kubenswrapper[4726]: I1001 16:23:10.764677 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-catalog-content\") pod \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\" (UID: \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\") " Oct 01 16:23:10 crc kubenswrapper[4726]: I1001 16:23:10.764840 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-utilities\") pod \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\" (UID: \"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a\") " Oct 01 16:23:10 crc kubenswrapper[4726]: I1001 16:23:10.765779 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-utilities" (OuterVolumeSpecName: "utilities") pod "06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" (UID: "06b06ea2-f3f3-45df-89a4-6a6a6d0c067a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:23:10 crc kubenswrapper[4726]: I1001 16:23:10.771826 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-kube-api-access-dm95l" (OuterVolumeSpecName: "kube-api-access-dm95l") pod "06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" (UID: "06b06ea2-f3f3-45df-89a4-6a6a6d0c067a"). InnerVolumeSpecName "kube-api-access-dm95l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:23:10 crc kubenswrapper[4726]: I1001 16:23:10.777967 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" (UID: "06b06ea2-f3f3-45df-89a4-6a6a6d0c067a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:23:10 crc kubenswrapper[4726]: I1001 16:23:10.867944 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dm95l\" (UniqueName: \"kubernetes.io/projected/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-kube-api-access-dm95l\") on node \"crc\" DevicePath \"\"" Oct 01 16:23:10 crc kubenswrapper[4726]: I1001 16:23:10.867980 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:23:10 crc kubenswrapper[4726]: I1001 16:23:10.867992 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.169508 4726 generic.go:334] "Generic (PLEG): container finished" podID="06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" containerID="e76ea193ba740035f09291335dc73d7cdb81e7d2a57d6d22f1b284a97bb05d63" exitCode=0 Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.169561 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jjmn4" event={"ID":"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a","Type":"ContainerDied","Data":"e76ea193ba740035f09291335dc73d7cdb81e7d2a57d6d22f1b284a97bb05d63"} Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.169592 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jjmn4" event={"ID":"06b06ea2-f3f3-45df-89a4-6a6a6d0c067a","Type":"ContainerDied","Data":"183bde35048fa7aafe7a91e86ef32a3c9fcda182e78398841b5874e485adb8b7"} Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.169612 4726 scope.go:117] "RemoveContainer" containerID="e76ea193ba740035f09291335dc73d7cdb81e7d2a57d6d22f1b284a97bb05d63" Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.169607 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jjmn4" Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.202338 4726 scope.go:117] "RemoveContainer" containerID="2eebe96a9fd303d3b364239f425d4a95158561869f1049eb1776d932ea9185ec" Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.210889 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jjmn4"] Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.222105 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jjmn4"] Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.233189 4726 scope.go:117] "RemoveContainer" containerID="85281e5a919e1685d3fee888b4da0426daa6d668cdbd64b12a2660c2f924e7c2" Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.264477 4726 scope.go:117] "RemoveContainer" containerID="e76ea193ba740035f09291335dc73d7cdb81e7d2a57d6d22f1b284a97bb05d63" Oct 01 16:23:11 crc kubenswrapper[4726]: E1001 16:23:11.267439 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e76ea193ba740035f09291335dc73d7cdb81e7d2a57d6d22f1b284a97bb05d63\": container with ID starting with e76ea193ba740035f09291335dc73d7cdb81e7d2a57d6d22f1b284a97bb05d63 not found: ID does not exist" containerID="e76ea193ba740035f09291335dc73d7cdb81e7d2a57d6d22f1b284a97bb05d63" Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.267483 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e76ea193ba740035f09291335dc73d7cdb81e7d2a57d6d22f1b284a97bb05d63"} err="failed to get container status \"e76ea193ba740035f09291335dc73d7cdb81e7d2a57d6d22f1b284a97bb05d63\": rpc error: code = NotFound desc = could not find container \"e76ea193ba740035f09291335dc73d7cdb81e7d2a57d6d22f1b284a97bb05d63\": container with ID starting with e76ea193ba740035f09291335dc73d7cdb81e7d2a57d6d22f1b284a97bb05d63 not found: ID does not exist" Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.267510 4726 scope.go:117] "RemoveContainer" containerID="2eebe96a9fd303d3b364239f425d4a95158561869f1049eb1776d932ea9185ec" Oct 01 16:23:11 crc kubenswrapper[4726]: E1001 16:23:11.267902 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eebe96a9fd303d3b364239f425d4a95158561869f1049eb1776d932ea9185ec\": container with ID starting with 2eebe96a9fd303d3b364239f425d4a95158561869f1049eb1776d932ea9185ec not found: ID does not exist" containerID="2eebe96a9fd303d3b364239f425d4a95158561869f1049eb1776d932ea9185ec" Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.267938 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eebe96a9fd303d3b364239f425d4a95158561869f1049eb1776d932ea9185ec"} err="failed to get container status \"2eebe96a9fd303d3b364239f425d4a95158561869f1049eb1776d932ea9185ec\": rpc error: code = NotFound desc = could not find container \"2eebe96a9fd303d3b364239f425d4a95158561869f1049eb1776d932ea9185ec\": container with ID starting with 2eebe96a9fd303d3b364239f425d4a95158561869f1049eb1776d932ea9185ec not found: ID does not exist" Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.267956 4726 scope.go:117] "RemoveContainer" containerID="85281e5a919e1685d3fee888b4da0426daa6d668cdbd64b12a2660c2f924e7c2" Oct 01 16:23:11 crc kubenswrapper[4726]: E1001 16:23:11.268500 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85281e5a919e1685d3fee888b4da0426daa6d668cdbd64b12a2660c2f924e7c2\": container with ID starting with 85281e5a919e1685d3fee888b4da0426daa6d668cdbd64b12a2660c2f924e7c2 not found: ID does not exist" containerID="85281e5a919e1685d3fee888b4da0426daa6d668cdbd64b12a2660c2f924e7c2" Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.268722 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85281e5a919e1685d3fee888b4da0426daa6d668cdbd64b12a2660c2f924e7c2"} err="failed to get container status \"85281e5a919e1685d3fee888b4da0426daa6d668cdbd64b12a2660c2f924e7c2\": rpc error: code = NotFound desc = could not find container \"85281e5a919e1685d3fee888b4da0426daa6d668cdbd64b12a2660c2f924e7c2\": container with ID starting with 85281e5a919e1685d3fee888b4da0426daa6d668cdbd64b12a2660c2f924e7c2 not found: ID does not exist" Oct 01 16:23:11 crc kubenswrapper[4726]: I1001 16:23:11.820864 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" path="/var/lib/kubelet/pods/06b06ea2-f3f3-45df-89a4-6a6a6d0c067a/volumes" Oct 01 16:23:21 crc kubenswrapper[4726]: I1001 16:23:21.373140 4726 scope.go:117] "RemoveContainer" containerID="900a97763a64952166995f3d2d97c0466fd11a3a256e59766c0bed6b95e2dadc" Oct 01 16:23:21 crc kubenswrapper[4726]: I1001 16:23:21.396930 4726 scope.go:117] "RemoveContainer" containerID="7d6aecbe07d7f7d01890bd04f0856e0f85bde44f6889434a4b4d2b78c0feed64" Oct 01 16:23:21 crc kubenswrapper[4726]: I1001 16:23:21.453416 4726 scope.go:117] "RemoveContainer" containerID="42be7ee4e31fdf3f8dd88cde48115931ef0a716d21e2846559fedede9c6f076a" Oct 01 16:23:38 crc kubenswrapper[4726]: I1001 16:23:38.838841 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v67xg"] Oct 01 16:23:38 crc kubenswrapper[4726]: E1001 16:23:38.839850 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" containerName="extract-utilities" Oct 01 16:23:38 crc kubenswrapper[4726]: I1001 16:23:38.839864 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" containerName="extract-utilities" Oct 01 16:23:38 crc kubenswrapper[4726]: E1001 16:23:38.839890 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" containerName="registry-server" Oct 01 16:23:38 crc kubenswrapper[4726]: I1001 16:23:38.839896 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" containerName="registry-server" Oct 01 16:23:38 crc kubenswrapper[4726]: E1001 16:23:38.839910 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" containerName="extract-content" Oct 01 16:23:38 crc kubenswrapper[4726]: I1001 16:23:38.839917 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" containerName="extract-content" Oct 01 16:23:38 crc kubenswrapper[4726]: I1001 16:23:38.840123 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="06b06ea2-f3f3-45df-89a4-6a6a6d0c067a" containerName="registry-server" Oct 01 16:23:38 crc kubenswrapper[4726]: I1001 16:23:38.841551 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:38 crc kubenswrapper[4726]: I1001 16:23:38.850838 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v67xg"] Oct 01 16:23:38 crc kubenswrapper[4726]: I1001 16:23:38.927571 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq89k\" (UniqueName: \"kubernetes.io/projected/ce9cb01d-d0b6-4fd8-90af-035396b9c425-kube-api-access-zq89k\") pod \"certified-operators-v67xg\" (UID: \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\") " pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:38 crc kubenswrapper[4726]: I1001 16:23:38.927683 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce9cb01d-d0b6-4fd8-90af-035396b9c425-catalog-content\") pod \"certified-operators-v67xg\" (UID: \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\") " pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:38 crc kubenswrapper[4726]: I1001 16:23:38.927764 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce9cb01d-d0b6-4fd8-90af-035396b9c425-utilities\") pod \"certified-operators-v67xg\" (UID: \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\") " pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:39 crc kubenswrapper[4726]: I1001 16:23:39.031036 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq89k\" (UniqueName: \"kubernetes.io/projected/ce9cb01d-d0b6-4fd8-90af-035396b9c425-kube-api-access-zq89k\") pod \"certified-operators-v67xg\" (UID: \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\") " pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:39 crc kubenswrapper[4726]: I1001 16:23:39.031229 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce9cb01d-d0b6-4fd8-90af-035396b9c425-catalog-content\") pod \"certified-operators-v67xg\" (UID: \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\") " pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:39 crc kubenswrapper[4726]: I1001 16:23:39.031348 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce9cb01d-d0b6-4fd8-90af-035396b9c425-utilities\") pod \"certified-operators-v67xg\" (UID: \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\") " pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:39 crc kubenswrapper[4726]: I1001 16:23:39.032039 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce9cb01d-d0b6-4fd8-90af-035396b9c425-utilities\") pod \"certified-operators-v67xg\" (UID: \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\") " pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:39 crc kubenswrapper[4726]: I1001 16:23:39.032671 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce9cb01d-d0b6-4fd8-90af-035396b9c425-catalog-content\") pod \"certified-operators-v67xg\" (UID: \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\") " pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:39 crc kubenswrapper[4726]: I1001 16:23:39.070957 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq89k\" (UniqueName: \"kubernetes.io/projected/ce9cb01d-d0b6-4fd8-90af-035396b9c425-kube-api-access-zq89k\") pod \"certified-operators-v67xg\" (UID: \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\") " pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:39 crc kubenswrapper[4726]: I1001 16:23:39.166917 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:39 crc kubenswrapper[4726]: I1001 16:23:39.676526 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v67xg"] Oct 01 16:23:40 crc kubenswrapper[4726]: I1001 16:23:40.485857 4726 generic.go:334] "Generic (PLEG): container finished" podID="ce9cb01d-d0b6-4fd8-90af-035396b9c425" containerID="534009e9cf31d12738e2881d74532669c335bbe9c5f48471e5df8220d34c9016" exitCode=0 Oct 01 16:23:40 crc kubenswrapper[4726]: I1001 16:23:40.486007 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v67xg" event={"ID":"ce9cb01d-d0b6-4fd8-90af-035396b9c425","Type":"ContainerDied","Data":"534009e9cf31d12738e2881d74532669c335bbe9c5f48471e5df8220d34c9016"} Oct 01 16:23:40 crc kubenswrapper[4726]: I1001 16:23:40.486290 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v67xg" event={"ID":"ce9cb01d-d0b6-4fd8-90af-035396b9c425","Type":"ContainerStarted","Data":"b9e9eefdd1c77e52cfee25f8caa048308eb88b8c2f5a52a02ea1f59114d69944"} Oct 01 16:23:41 crc kubenswrapper[4726]: I1001 16:23:41.499179 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v67xg" event={"ID":"ce9cb01d-d0b6-4fd8-90af-035396b9c425","Type":"ContainerStarted","Data":"b0325ebd8595901e81594eccac936c9ebd7da602cc931d7cf21f5ba28b0e9256"} Oct 01 16:23:43 crc kubenswrapper[4726]: I1001 16:23:43.529352 4726 generic.go:334] "Generic (PLEG): container finished" podID="ce9cb01d-d0b6-4fd8-90af-035396b9c425" containerID="b0325ebd8595901e81594eccac936c9ebd7da602cc931d7cf21f5ba28b0e9256" exitCode=0 Oct 01 16:23:43 crc kubenswrapper[4726]: I1001 16:23:43.529426 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v67xg" event={"ID":"ce9cb01d-d0b6-4fd8-90af-035396b9c425","Type":"ContainerDied","Data":"b0325ebd8595901e81594eccac936c9ebd7da602cc931d7cf21f5ba28b0e9256"} Oct 01 16:23:44 crc kubenswrapper[4726]: I1001 16:23:44.540240 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v67xg" event={"ID":"ce9cb01d-d0b6-4fd8-90af-035396b9c425","Type":"ContainerStarted","Data":"60f38525498b6c64e1c180603f9b478f3a180e7a3ca6f7fcb0ae59f057104eb7"} Oct 01 16:23:44 crc kubenswrapper[4726]: I1001 16:23:44.571427 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v67xg" podStartSLOduration=3.008843947 podStartE2EDuration="6.571407212s" podCreationTimestamp="2025-10-01 16:23:38 +0000 UTC" firstStartedPulling="2025-10-01 16:23:40.488040193 +0000 UTC m=+1473.389592770" lastFinishedPulling="2025-10-01 16:23:44.050603468 +0000 UTC m=+1476.952156035" observedRunningTime="2025-10-01 16:23:44.563226879 +0000 UTC m=+1477.464779456" watchObservedRunningTime="2025-10-01 16:23:44.571407212 +0000 UTC m=+1477.472959789" Oct 01 16:23:49 crc kubenswrapper[4726]: I1001 16:23:49.167299 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:49 crc kubenswrapper[4726]: I1001 16:23:49.167873 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:49 crc kubenswrapper[4726]: I1001 16:23:49.221909 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:49 crc kubenswrapper[4726]: I1001 16:23:49.642812 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:49 crc kubenswrapper[4726]: I1001 16:23:49.698037 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v67xg"] Oct 01 16:23:51 crc kubenswrapper[4726]: I1001 16:23:51.612157 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v67xg" podUID="ce9cb01d-d0b6-4fd8-90af-035396b9c425" containerName="registry-server" containerID="cri-o://60f38525498b6c64e1c180603f9b478f3a180e7a3ca6f7fcb0ae59f057104eb7" gracePeriod=2 Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.132828 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.207090 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce9cb01d-d0b6-4fd8-90af-035396b9c425-catalog-content\") pod \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\" (UID: \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\") " Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.207213 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zq89k\" (UniqueName: \"kubernetes.io/projected/ce9cb01d-d0b6-4fd8-90af-035396b9c425-kube-api-access-zq89k\") pod \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\" (UID: \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\") " Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.207291 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce9cb01d-d0b6-4fd8-90af-035396b9c425-utilities\") pod \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\" (UID: \"ce9cb01d-d0b6-4fd8-90af-035396b9c425\") " Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.208276 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce9cb01d-d0b6-4fd8-90af-035396b9c425-utilities" (OuterVolumeSpecName: "utilities") pod "ce9cb01d-d0b6-4fd8-90af-035396b9c425" (UID: "ce9cb01d-d0b6-4fd8-90af-035396b9c425"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.220398 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce9cb01d-d0b6-4fd8-90af-035396b9c425-kube-api-access-zq89k" (OuterVolumeSpecName: "kube-api-access-zq89k") pod "ce9cb01d-d0b6-4fd8-90af-035396b9c425" (UID: "ce9cb01d-d0b6-4fd8-90af-035396b9c425"). InnerVolumeSpecName "kube-api-access-zq89k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.253324 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce9cb01d-d0b6-4fd8-90af-035396b9c425-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce9cb01d-d0b6-4fd8-90af-035396b9c425" (UID: "ce9cb01d-d0b6-4fd8-90af-035396b9c425"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.310207 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zq89k\" (UniqueName: \"kubernetes.io/projected/ce9cb01d-d0b6-4fd8-90af-035396b9c425-kube-api-access-zq89k\") on node \"crc\" DevicePath \"\"" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.310253 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce9cb01d-d0b6-4fd8-90af-035396b9c425-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.310269 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce9cb01d-d0b6-4fd8-90af-035396b9c425-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.626002 4726 generic.go:334] "Generic (PLEG): container finished" podID="ce9cb01d-d0b6-4fd8-90af-035396b9c425" containerID="60f38525498b6c64e1c180603f9b478f3a180e7a3ca6f7fcb0ae59f057104eb7" exitCode=0 Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.626083 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v67xg" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.626075 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v67xg" event={"ID":"ce9cb01d-d0b6-4fd8-90af-035396b9c425","Type":"ContainerDied","Data":"60f38525498b6c64e1c180603f9b478f3a180e7a3ca6f7fcb0ae59f057104eb7"} Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.627401 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v67xg" event={"ID":"ce9cb01d-d0b6-4fd8-90af-035396b9c425","Type":"ContainerDied","Data":"b9e9eefdd1c77e52cfee25f8caa048308eb88b8c2f5a52a02ea1f59114d69944"} Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.627424 4726 scope.go:117] "RemoveContainer" containerID="60f38525498b6c64e1c180603f9b478f3a180e7a3ca6f7fcb0ae59f057104eb7" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.657661 4726 scope.go:117] "RemoveContainer" containerID="b0325ebd8595901e81594eccac936c9ebd7da602cc931d7cf21f5ba28b0e9256" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.662574 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v67xg"] Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.674448 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v67xg"] Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.698079 4726 scope.go:117] "RemoveContainer" containerID="534009e9cf31d12738e2881d74532669c335bbe9c5f48471e5df8220d34c9016" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.725407 4726 scope.go:117] "RemoveContainer" containerID="60f38525498b6c64e1c180603f9b478f3a180e7a3ca6f7fcb0ae59f057104eb7" Oct 01 16:23:52 crc kubenswrapper[4726]: E1001 16:23:52.725704 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60f38525498b6c64e1c180603f9b478f3a180e7a3ca6f7fcb0ae59f057104eb7\": container with ID starting with 60f38525498b6c64e1c180603f9b478f3a180e7a3ca6f7fcb0ae59f057104eb7 not found: ID does not exist" containerID="60f38525498b6c64e1c180603f9b478f3a180e7a3ca6f7fcb0ae59f057104eb7" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.725734 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60f38525498b6c64e1c180603f9b478f3a180e7a3ca6f7fcb0ae59f057104eb7"} err="failed to get container status \"60f38525498b6c64e1c180603f9b478f3a180e7a3ca6f7fcb0ae59f057104eb7\": rpc error: code = NotFound desc = could not find container \"60f38525498b6c64e1c180603f9b478f3a180e7a3ca6f7fcb0ae59f057104eb7\": container with ID starting with 60f38525498b6c64e1c180603f9b478f3a180e7a3ca6f7fcb0ae59f057104eb7 not found: ID does not exist" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.725754 4726 scope.go:117] "RemoveContainer" containerID="b0325ebd8595901e81594eccac936c9ebd7da602cc931d7cf21f5ba28b0e9256" Oct 01 16:23:52 crc kubenswrapper[4726]: E1001 16:23:52.726105 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0325ebd8595901e81594eccac936c9ebd7da602cc931d7cf21f5ba28b0e9256\": container with ID starting with b0325ebd8595901e81594eccac936c9ebd7da602cc931d7cf21f5ba28b0e9256 not found: ID does not exist" containerID="b0325ebd8595901e81594eccac936c9ebd7da602cc931d7cf21f5ba28b0e9256" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.726127 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0325ebd8595901e81594eccac936c9ebd7da602cc931d7cf21f5ba28b0e9256"} err="failed to get container status \"b0325ebd8595901e81594eccac936c9ebd7da602cc931d7cf21f5ba28b0e9256\": rpc error: code = NotFound desc = could not find container \"b0325ebd8595901e81594eccac936c9ebd7da602cc931d7cf21f5ba28b0e9256\": container with ID starting with b0325ebd8595901e81594eccac936c9ebd7da602cc931d7cf21f5ba28b0e9256 not found: ID does not exist" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.726141 4726 scope.go:117] "RemoveContainer" containerID="534009e9cf31d12738e2881d74532669c335bbe9c5f48471e5df8220d34c9016" Oct 01 16:23:52 crc kubenswrapper[4726]: E1001 16:23:52.726454 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"534009e9cf31d12738e2881d74532669c335bbe9c5f48471e5df8220d34c9016\": container with ID starting with 534009e9cf31d12738e2881d74532669c335bbe9c5f48471e5df8220d34c9016 not found: ID does not exist" containerID="534009e9cf31d12738e2881d74532669c335bbe9c5f48471e5df8220d34c9016" Oct 01 16:23:52 crc kubenswrapper[4726]: I1001 16:23:52.726566 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"534009e9cf31d12738e2881d74532669c335bbe9c5f48471e5df8220d34c9016"} err="failed to get container status \"534009e9cf31d12738e2881d74532669c335bbe9c5f48471e5df8220d34c9016\": rpc error: code = NotFound desc = could not find container \"534009e9cf31d12738e2881d74532669c335bbe9c5f48471e5df8220d34c9016\": container with ID starting with 534009e9cf31d12738e2881d74532669c335bbe9c5f48471e5df8220d34c9016 not found: ID does not exist" Oct 01 16:23:53 crc kubenswrapper[4726]: I1001 16:23:53.817575 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce9cb01d-d0b6-4fd8-90af-035396b9c425" path="/var/lib/kubelet/pods/ce9cb01d-d0b6-4fd8-90af-035396b9c425/volumes" Oct 01 16:24:23 crc kubenswrapper[4726]: I1001 16:24:23.413446 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:24:23 crc kubenswrapper[4726]: I1001 16:24:23.414020 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:24:53 crc kubenswrapper[4726]: I1001 16:24:53.414691 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:24:53 crc kubenswrapper[4726]: I1001 16:24:53.416038 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.413860 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.414571 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.414624 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.415329 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.415386 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" gracePeriod=600 Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.583683 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" exitCode=0 Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.583724 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840"} Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.583754 4726 scope.go:117] "RemoveContainer" containerID="f2fff8833153004baace8bca4e9e33e62ef29d5bce54ddbea47fab14e44b1073" Oct 01 16:25:23 crc kubenswrapper[4726]: E1001 16:25:23.611467 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.883916 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-km6h4"] Oct 01 16:25:23 crc kubenswrapper[4726]: E1001 16:25:23.884382 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce9cb01d-d0b6-4fd8-90af-035396b9c425" containerName="extract-utilities" Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.884402 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce9cb01d-d0b6-4fd8-90af-035396b9c425" containerName="extract-utilities" Oct 01 16:25:23 crc kubenswrapper[4726]: E1001 16:25:23.884432 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce9cb01d-d0b6-4fd8-90af-035396b9c425" containerName="extract-content" Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.884441 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce9cb01d-d0b6-4fd8-90af-035396b9c425" containerName="extract-content" Oct 01 16:25:23 crc kubenswrapper[4726]: E1001 16:25:23.884455 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce9cb01d-d0b6-4fd8-90af-035396b9c425" containerName="registry-server" Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.884463 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce9cb01d-d0b6-4fd8-90af-035396b9c425" containerName="registry-server" Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.884678 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce9cb01d-d0b6-4fd8-90af-035396b9c425" containerName="registry-server" Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.886646 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.903585 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-km6h4"] Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.998839 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/299aa258-7b3f-4759-bfc2-abeecf066d81-utilities\") pod \"community-operators-km6h4\" (UID: \"299aa258-7b3f-4759-bfc2-abeecf066d81\") " pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.998928 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/299aa258-7b3f-4759-bfc2-abeecf066d81-catalog-content\") pod \"community-operators-km6h4\" (UID: \"299aa258-7b3f-4759-bfc2-abeecf066d81\") " pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:23 crc kubenswrapper[4726]: I1001 16:25:23.999031 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9lc4\" (UniqueName: \"kubernetes.io/projected/299aa258-7b3f-4759-bfc2-abeecf066d81-kube-api-access-d9lc4\") pod \"community-operators-km6h4\" (UID: \"299aa258-7b3f-4759-bfc2-abeecf066d81\") " pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:24 crc kubenswrapper[4726]: I1001 16:25:24.101178 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/299aa258-7b3f-4759-bfc2-abeecf066d81-utilities\") pod \"community-operators-km6h4\" (UID: \"299aa258-7b3f-4759-bfc2-abeecf066d81\") " pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:24 crc kubenswrapper[4726]: I1001 16:25:24.101258 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/299aa258-7b3f-4759-bfc2-abeecf066d81-catalog-content\") pod \"community-operators-km6h4\" (UID: \"299aa258-7b3f-4759-bfc2-abeecf066d81\") " pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:24 crc kubenswrapper[4726]: I1001 16:25:24.101403 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9lc4\" (UniqueName: \"kubernetes.io/projected/299aa258-7b3f-4759-bfc2-abeecf066d81-kube-api-access-d9lc4\") pod \"community-operators-km6h4\" (UID: \"299aa258-7b3f-4759-bfc2-abeecf066d81\") " pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:24 crc kubenswrapper[4726]: I1001 16:25:24.101863 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/299aa258-7b3f-4759-bfc2-abeecf066d81-utilities\") pod \"community-operators-km6h4\" (UID: \"299aa258-7b3f-4759-bfc2-abeecf066d81\") " pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:24 crc kubenswrapper[4726]: I1001 16:25:24.101936 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/299aa258-7b3f-4759-bfc2-abeecf066d81-catalog-content\") pod \"community-operators-km6h4\" (UID: \"299aa258-7b3f-4759-bfc2-abeecf066d81\") " pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:24 crc kubenswrapper[4726]: I1001 16:25:24.125798 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9lc4\" (UniqueName: \"kubernetes.io/projected/299aa258-7b3f-4759-bfc2-abeecf066d81-kube-api-access-d9lc4\") pod \"community-operators-km6h4\" (UID: \"299aa258-7b3f-4759-bfc2-abeecf066d81\") " pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:24 crc kubenswrapper[4726]: I1001 16:25:24.209473 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:24 crc kubenswrapper[4726]: I1001 16:25:24.595494 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:25:24 crc kubenswrapper[4726]: E1001 16:25:24.596126 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:25:24 crc kubenswrapper[4726]: I1001 16:25:24.763059 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-km6h4"] Oct 01 16:25:25 crc kubenswrapper[4726]: I1001 16:25:25.607369 4726 generic.go:334] "Generic (PLEG): container finished" podID="299aa258-7b3f-4759-bfc2-abeecf066d81" containerID="b5a0a6d69e11f01134fdba6fe625af0444845d3d71742945a7a78af3d564bdb8" exitCode=0 Oct 01 16:25:25 crc kubenswrapper[4726]: I1001 16:25:25.607481 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-km6h4" event={"ID":"299aa258-7b3f-4759-bfc2-abeecf066d81","Type":"ContainerDied","Data":"b5a0a6d69e11f01134fdba6fe625af0444845d3d71742945a7a78af3d564bdb8"} Oct 01 16:25:25 crc kubenswrapper[4726]: I1001 16:25:25.607744 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-km6h4" event={"ID":"299aa258-7b3f-4759-bfc2-abeecf066d81","Type":"ContainerStarted","Data":"2cc8b4fa18a5720136d4dfe36868c30984326b74b21202c0ef8225ad9c1d6df7"} Oct 01 16:25:25 crc kubenswrapper[4726]: I1001 16:25:25.609420 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:25:27 crc kubenswrapper[4726]: I1001 16:25:27.629601 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-km6h4" event={"ID":"299aa258-7b3f-4759-bfc2-abeecf066d81","Type":"ContainerStarted","Data":"ee3148184d5ccd2789d2ffff21570bdd00080225cb7b00cfc6d2a56d89ca4ad1"} Oct 01 16:25:28 crc kubenswrapper[4726]: I1001 16:25:28.640877 4726 generic.go:334] "Generic (PLEG): container finished" podID="299aa258-7b3f-4759-bfc2-abeecf066d81" containerID="ee3148184d5ccd2789d2ffff21570bdd00080225cb7b00cfc6d2a56d89ca4ad1" exitCode=0 Oct 01 16:25:28 crc kubenswrapper[4726]: I1001 16:25:28.640969 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-km6h4" event={"ID":"299aa258-7b3f-4759-bfc2-abeecf066d81","Type":"ContainerDied","Data":"ee3148184d5ccd2789d2ffff21570bdd00080225cb7b00cfc6d2a56d89ca4ad1"} Oct 01 16:25:29 crc kubenswrapper[4726]: I1001 16:25:29.659342 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-km6h4" event={"ID":"299aa258-7b3f-4759-bfc2-abeecf066d81","Type":"ContainerStarted","Data":"e0b7828c5a96d9595adf657bc6d00a771fa80b7937f6b4ce6d9ad94831f7583a"} Oct 01 16:25:29 crc kubenswrapper[4726]: I1001 16:25:29.677387 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-km6h4" podStartSLOduration=3.175300393 podStartE2EDuration="6.67737096s" podCreationTimestamp="2025-10-01 16:25:23 +0000 UTC" firstStartedPulling="2025-10-01 16:25:25.609205152 +0000 UTC m=+1578.510757729" lastFinishedPulling="2025-10-01 16:25:29.111275719 +0000 UTC m=+1582.012828296" observedRunningTime="2025-10-01 16:25:29.676719621 +0000 UTC m=+1582.578272208" watchObservedRunningTime="2025-10-01 16:25:29.67737096 +0000 UTC m=+1582.578923527" Oct 01 16:25:34 crc kubenswrapper[4726]: I1001 16:25:34.209675 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:34 crc kubenswrapper[4726]: I1001 16:25:34.210226 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:34 crc kubenswrapper[4726]: I1001 16:25:34.311477 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:34 crc kubenswrapper[4726]: I1001 16:25:34.714022 4726 generic.go:334] "Generic (PLEG): container finished" podID="bf9f4ecc-09d0-482c-81d2-d7fe9503de1f" containerID="1d4bce1a8d60503608b9e886c2ef531096c709b261cec5f703135bef8185395a" exitCode=0 Oct 01 16:25:34 crc kubenswrapper[4726]: I1001 16:25:34.714102 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" event={"ID":"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f","Type":"ContainerDied","Data":"1d4bce1a8d60503608b9e886c2ef531096c709b261cec5f703135bef8185395a"} Oct 01 16:25:34 crc kubenswrapper[4726]: I1001 16:25:34.777843 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:34 crc kubenswrapper[4726]: I1001 16:25:34.819185 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-km6h4"] Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.152506 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.157862 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwfnm\" (UniqueName: \"kubernetes.io/projected/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-kube-api-access-jwfnm\") pod \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.157989 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-ssh-key\") pod \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.158089 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-inventory\") pod \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.159313 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-bootstrap-combined-ca-bundle\") pod \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\" (UID: \"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f\") " Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.165894 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-kube-api-access-jwfnm" (OuterVolumeSpecName: "kube-api-access-jwfnm") pod "bf9f4ecc-09d0-482c-81d2-d7fe9503de1f" (UID: "bf9f4ecc-09d0-482c-81d2-d7fe9503de1f"). InnerVolumeSpecName "kube-api-access-jwfnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.168200 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "bf9f4ecc-09d0-482c-81d2-d7fe9503de1f" (UID: "bf9f4ecc-09d0-482c-81d2-d7fe9503de1f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.196069 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-inventory" (OuterVolumeSpecName: "inventory") pod "bf9f4ecc-09d0-482c-81d2-d7fe9503de1f" (UID: "bf9f4ecc-09d0-482c-81d2-d7fe9503de1f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.206188 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bf9f4ecc-09d0-482c-81d2-d7fe9503de1f" (UID: "bf9f4ecc-09d0-482c-81d2-d7fe9503de1f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.261604 4726 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.261765 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwfnm\" (UniqueName: \"kubernetes.io/projected/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-kube-api-access-jwfnm\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.261785 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.261808 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9f4ecc-09d0-482c-81d2-d7fe9503de1f-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.735222 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-km6h4" podUID="299aa258-7b3f-4759-bfc2-abeecf066d81" containerName="registry-server" containerID="cri-o://e0b7828c5a96d9595adf657bc6d00a771fa80b7937f6b4ce6d9ad94831f7583a" gracePeriod=2 Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.735275 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.735302 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd" event={"ID":"bf9f4ecc-09d0-482c-81d2-d7fe9503de1f","Type":"ContainerDied","Data":"c2e519cfc09459945a82c96c3f1753582dd290313e93e7bba57614344f38c469"} Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.736091 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2e519cfc09459945a82c96c3f1753582dd290313e93e7bba57614344f38c469" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.808649 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:25:36 crc kubenswrapper[4726]: E1001 16:25:36.809092 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.836289 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58"] Oct 01 16:25:36 crc kubenswrapper[4726]: E1001 16:25:36.836781 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf9f4ecc-09d0-482c-81d2-d7fe9503de1f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.836799 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf9f4ecc-09d0-482c-81d2-d7fe9503de1f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.837028 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf9f4ecc-09d0-482c-81d2-d7fe9503de1f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.837814 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.846462 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.846907 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.847325 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.848665 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.850427 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58"] Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.996737 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5rs4\" (UniqueName: \"kubernetes.io/projected/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-kube-api-access-c5rs4\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-g7c58\" (UID: \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.996908 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-g7c58\" (UID: \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" Oct 01 16:25:36 crc kubenswrapper[4726]: I1001 16:25:36.997418 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-g7c58\" (UID: \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.101168 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-g7c58\" (UID: \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.101596 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5rs4\" (UniqueName: \"kubernetes.io/projected/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-kube-api-access-c5rs4\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-g7c58\" (UID: \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.101642 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-g7c58\" (UID: \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.107693 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-g7c58\" (UID: \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.108120 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-g7c58\" (UID: \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.126360 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5rs4\" (UniqueName: \"kubernetes.io/projected/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-kube-api-access-c5rs4\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-g7c58\" (UID: \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.229244 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.265181 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.307648 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9lc4\" (UniqueName: \"kubernetes.io/projected/299aa258-7b3f-4759-bfc2-abeecf066d81-kube-api-access-d9lc4\") pod \"299aa258-7b3f-4759-bfc2-abeecf066d81\" (UID: \"299aa258-7b3f-4759-bfc2-abeecf066d81\") " Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.307795 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/299aa258-7b3f-4759-bfc2-abeecf066d81-utilities\") pod \"299aa258-7b3f-4759-bfc2-abeecf066d81\" (UID: \"299aa258-7b3f-4759-bfc2-abeecf066d81\") " Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.307989 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/299aa258-7b3f-4759-bfc2-abeecf066d81-catalog-content\") pod \"299aa258-7b3f-4759-bfc2-abeecf066d81\" (UID: \"299aa258-7b3f-4759-bfc2-abeecf066d81\") " Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.309020 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/299aa258-7b3f-4759-bfc2-abeecf066d81-utilities" (OuterVolumeSpecName: "utilities") pod "299aa258-7b3f-4759-bfc2-abeecf066d81" (UID: "299aa258-7b3f-4759-bfc2-abeecf066d81"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.315483 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/299aa258-7b3f-4759-bfc2-abeecf066d81-kube-api-access-d9lc4" (OuterVolumeSpecName: "kube-api-access-d9lc4") pod "299aa258-7b3f-4759-bfc2-abeecf066d81" (UID: "299aa258-7b3f-4759-bfc2-abeecf066d81"). InnerVolumeSpecName "kube-api-access-d9lc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.394624 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/299aa258-7b3f-4759-bfc2-abeecf066d81-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "299aa258-7b3f-4759-bfc2-abeecf066d81" (UID: "299aa258-7b3f-4759-bfc2-abeecf066d81"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.409783 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/299aa258-7b3f-4759-bfc2-abeecf066d81-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.409818 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/299aa258-7b3f-4759-bfc2-abeecf066d81-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.409831 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9lc4\" (UniqueName: \"kubernetes.io/projected/299aa258-7b3f-4759-bfc2-abeecf066d81-kube-api-access-d9lc4\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.751668 4726 generic.go:334] "Generic (PLEG): container finished" podID="299aa258-7b3f-4759-bfc2-abeecf066d81" containerID="e0b7828c5a96d9595adf657bc6d00a771fa80b7937f6b4ce6d9ad94831f7583a" exitCode=0 Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.751895 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-km6h4" event={"ID":"299aa258-7b3f-4759-bfc2-abeecf066d81","Type":"ContainerDied","Data":"e0b7828c5a96d9595adf657bc6d00a771fa80b7937f6b4ce6d9ad94831f7583a"} Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.752005 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-km6h4" event={"ID":"299aa258-7b3f-4759-bfc2-abeecf066d81","Type":"ContainerDied","Data":"2cc8b4fa18a5720136d4dfe36868c30984326b74b21202c0ef8225ad9c1d6df7"} Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.752032 4726 scope.go:117] "RemoveContainer" containerID="e0b7828c5a96d9595adf657bc6d00a771fa80b7937f6b4ce6d9ad94831f7583a" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.752091 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-km6h4" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.782742 4726 scope.go:117] "RemoveContainer" containerID="ee3148184d5ccd2789d2ffff21570bdd00080225cb7b00cfc6d2a56d89ca4ad1" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.796374 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-km6h4"] Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.804803 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-km6h4"] Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.813539 4726 scope.go:117] "RemoveContainer" containerID="b5a0a6d69e11f01134fdba6fe625af0444845d3d71742945a7a78af3d564bdb8" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.828846 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="299aa258-7b3f-4759-bfc2-abeecf066d81" path="/var/lib/kubelet/pods/299aa258-7b3f-4759-bfc2-abeecf066d81/volumes" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.829770 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58"] Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.861459 4726 scope.go:117] "RemoveContainer" containerID="e0b7828c5a96d9595adf657bc6d00a771fa80b7937f6b4ce6d9ad94831f7583a" Oct 01 16:25:37 crc kubenswrapper[4726]: E1001 16:25:37.862184 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0b7828c5a96d9595adf657bc6d00a771fa80b7937f6b4ce6d9ad94831f7583a\": container with ID starting with e0b7828c5a96d9595adf657bc6d00a771fa80b7937f6b4ce6d9ad94831f7583a not found: ID does not exist" containerID="e0b7828c5a96d9595adf657bc6d00a771fa80b7937f6b4ce6d9ad94831f7583a" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.862234 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0b7828c5a96d9595adf657bc6d00a771fa80b7937f6b4ce6d9ad94831f7583a"} err="failed to get container status \"e0b7828c5a96d9595adf657bc6d00a771fa80b7937f6b4ce6d9ad94831f7583a\": rpc error: code = NotFound desc = could not find container \"e0b7828c5a96d9595adf657bc6d00a771fa80b7937f6b4ce6d9ad94831f7583a\": container with ID starting with e0b7828c5a96d9595adf657bc6d00a771fa80b7937f6b4ce6d9ad94831f7583a not found: ID does not exist" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.862261 4726 scope.go:117] "RemoveContainer" containerID="ee3148184d5ccd2789d2ffff21570bdd00080225cb7b00cfc6d2a56d89ca4ad1" Oct 01 16:25:37 crc kubenswrapper[4726]: E1001 16:25:37.862878 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee3148184d5ccd2789d2ffff21570bdd00080225cb7b00cfc6d2a56d89ca4ad1\": container with ID starting with ee3148184d5ccd2789d2ffff21570bdd00080225cb7b00cfc6d2a56d89ca4ad1 not found: ID does not exist" containerID="ee3148184d5ccd2789d2ffff21570bdd00080225cb7b00cfc6d2a56d89ca4ad1" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.862905 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee3148184d5ccd2789d2ffff21570bdd00080225cb7b00cfc6d2a56d89ca4ad1"} err="failed to get container status \"ee3148184d5ccd2789d2ffff21570bdd00080225cb7b00cfc6d2a56d89ca4ad1\": rpc error: code = NotFound desc = could not find container \"ee3148184d5ccd2789d2ffff21570bdd00080225cb7b00cfc6d2a56d89ca4ad1\": container with ID starting with ee3148184d5ccd2789d2ffff21570bdd00080225cb7b00cfc6d2a56d89ca4ad1 not found: ID does not exist" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.862924 4726 scope.go:117] "RemoveContainer" containerID="b5a0a6d69e11f01134fdba6fe625af0444845d3d71742945a7a78af3d564bdb8" Oct 01 16:25:37 crc kubenswrapper[4726]: E1001 16:25:37.863166 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5a0a6d69e11f01134fdba6fe625af0444845d3d71742945a7a78af3d564bdb8\": container with ID starting with b5a0a6d69e11f01134fdba6fe625af0444845d3d71742945a7a78af3d564bdb8 not found: ID does not exist" containerID="b5a0a6d69e11f01134fdba6fe625af0444845d3d71742945a7a78af3d564bdb8" Oct 01 16:25:37 crc kubenswrapper[4726]: I1001 16:25:37.863183 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5a0a6d69e11f01134fdba6fe625af0444845d3d71742945a7a78af3d564bdb8"} err="failed to get container status \"b5a0a6d69e11f01134fdba6fe625af0444845d3d71742945a7a78af3d564bdb8\": rpc error: code = NotFound desc = could not find container \"b5a0a6d69e11f01134fdba6fe625af0444845d3d71742945a7a78af3d564bdb8\": container with ID starting with b5a0a6d69e11f01134fdba6fe625af0444845d3d71742945a7a78af3d564bdb8 not found: ID does not exist" Oct 01 16:25:38 crc kubenswrapper[4726]: I1001 16:25:38.767367 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" event={"ID":"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b","Type":"ContainerStarted","Data":"dee6640eb5b49790ea972e7db284ffb2ce6aa1488f05a77f0c45bb4aceecad32"} Oct 01 16:25:38 crc kubenswrapper[4726]: I1001 16:25:38.767694 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" event={"ID":"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b","Type":"ContainerStarted","Data":"61008bdb23060b1b3876491ac32503768685bb89710f988a531694567cbfa49d"} Oct 01 16:25:38 crc kubenswrapper[4726]: I1001 16:25:38.798154 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" podStartSLOduration=2.312255388 podStartE2EDuration="2.798118851s" podCreationTimestamp="2025-10-01 16:25:36 +0000 UTC" firstStartedPulling="2025-10-01 16:25:37.82439986 +0000 UTC m=+1590.725952437" lastFinishedPulling="2025-10-01 16:25:38.310263323 +0000 UTC m=+1591.211815900" observedRunningTime="2025-10-01 16:25:38.788257681 +0000 UTC m=+1591.689810278" watchObservedRunningTime="2025-10-01 16:25:38.798118851 +0000 UTC m=+1591.699671438" Oct 01 16:25:41 crc kubenswrapper[4726]: I1001 16:25:41.043555 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-gcts2"] Oct 01 16:25:41 crc kubenswrapper[4726]: I1001 16:25:41.057372 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-b62tk"] Oct 01 16:25:41 crc kubenswrapper[4726]: I1001 16:25:41.065777 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-gcts2"] Oct 01 16:25:41 crc kubenswrapper[4726]: I1001 16:25:41.074506 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-b62tk"] Oct 01 16:25:41 crc kubenswrapper[4726]: I1001 16:25:41.822162 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15a72138-dafd-48f5-95cb-420e35e00609" path="/var/lib/kubelet/pods/15a72138-dafd-48f5-95cb-420e35e00609/volumes" Oct 01 16:25:41 crc kubenswrapper[4726]: I1001 16:25:41.823539 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e34fc66-c06d-4ed5-96f1-f54dc1fe4765" path="/var/lib/kubelet/pods/3e34fc66-c06d-4ed5-96f1-f54dc1fe4765/volumes" Oct 01 16:25:46 crc kubenswrapper[4726]: I1001 16:25:46.030998 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-gj94c"] Oct 01 16:25:46 crc kubenswrapper[4726]: I1001 16:25:46.038554 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-gj94c"] Oct 01 16:25:47 crc kubenswrapper[4726]: I1001 16:25:47.822233 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de31038e-fe8f-4d23-b875-aad3c596fbea" path="/var/lib/kubelet/pods/de31038e-fe8f-4d23-b875-aad3c596fbea/volumes" Oct 01 16:25:49 crc kubenswrapper[4726]: I1001 16:25:49.030400 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-m6p6g"] Oct 01 16:25:49 crc kubenswrapper[4726]: I1001 16:25:49.040401 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-k9qcv"] Oct 01 16:25:49 crc kubenswrapper[4726]: I1001 16:25:49.053661 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-m6p6g"] Oct 01 16:25:49 crc kubenswrapper[4726]: I1001 16:25:49.062844 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-z6kcl"] Oct 01 16:25:49 crc kubenswrapper[4726]: I1001 16:25:49.071272 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-k9qcv"] Oct 01 16:25:49 crc kubenswrapper[4726]: I1001 16:25:49.080384 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-z6kcl"] Oct 01 16:25:49 crc kubenswrapper[4726]: I1001 16:25:49.822996 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ca84aaa-d132-4716-9ec0-a6ccbf812ade" path="/var/lib/kubelet/pods/5ca84aaa-d132-4716-9ec0-a6ccbf812ade/volumes" Oct 01 16:25:49 crc kubenswrapper[4726]: I1001 16:25:49.823777 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="679126d8-15ab-4d02-bafe-72f753a5d526" path="/var/lib/kubelet/pods/679126d8-15ab-4d02-bafe-72f753a5d526/volumes" Oct 01 16:25:49 crc kubenswrapper[4726]: I1001 16:25:49.824409 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffa2e20c-77c7-436d-a548-f06977e9ae3e" path="/var/lib/kubelet/pods/ffa2e20c-77c7-436d-a548-f06977e9ae3e/volumes" Oct 01 16:25:51 crc kubenswrapper[4726]: I1001 16:25:51.031419 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9bb7-account-create-xzkf9"] Oct 01 16:25:51 crc kubenswrapper[4726]: I1001 16:25:51.041206 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-9457-account-create-6wcqq"] Oct 01 16:25:51 crc kubenswrapper[4726]: I1001 16:25:51.051437 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-9457-account-create-6wcqq"] Oct 01 16:25:51 crc kubenswrapper[4726]: I1001 16:25:51.060704 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9bb7-account-create-xzkf9"] Oct 01 16:25:51 crc kubenswrapper[4726]: I1001 16:25:51.810183 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:25:51 crc kubenswrapper[4726]: E1001 16:25:51.814277 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:25:51 crc kubenswrapper[4726]: I1001 16:25:51.828179 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68b428fe-6f95-4fbd-b31a-b3aafba90d32" path="/var/lib/kubelet/pods/68b428fe-6f95-4fbd-b31a-b3aafba90d32/volumes" Oct 01 16:25:51 crc kubenswrapper[4726]: I1001 16:25:51.828706 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b9a5008-3aed-4f59-9ad5-3f568f35c0dc" path="/var/lib/kubelet/pods/6b9a5008-3aed-4f59-9ad5-3f568f35c0dc/volumes" Oct 01 16:25:56 crc kubenswrapper[4726]: I1001 16:25:56.034462 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-6f33-account-create-rw9w7"] Oct 01 16:25:56 crc kubenswrapper[4726]: I1001 16:25:56.045041 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-6f33-account-create-rw9w7"] Oct 01 16:25:57 crc kubenswrapper[4726]: I1001 16:25:57.829690 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50c724bf-9bf3-4e8b-97c5-09199a794376" path="/var/lib/kubelet/pods/50c724bf-9bf3-4e8b-97c5-09199a794376/volumes" Oct 01 16:26:04 crc kubenswrapper[4726]: I1001 16:26:04.807572 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:26:04 crc kubenswrapper[4726]: E1001 16:26:04.809834 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:26:09 crc kubenswrapper[4726]: I1001 16:26:09.047955 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-7b0c-account-create-jf4fq"] Oct 01 16:26:09 crc kubenswrapper[4726]: I1001 16:26:09.060845 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-7b0c-account-create-jf4fq"] Oct 01 16:26:09 crc kubenswrapper[4726]: I1001 16:26:09.827344 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56bde538-7fa3-4a9b-be9f-87ee018e6a43" path="/var/lib/kubelet/pods/56bde538-7fa3-4a9b-be9f-87ee018e6a43/volumes" Oct 01 16:26:10 crc kubenswrapper[4726]: I1001 16:26:10.026970 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-6bf9-account-create-9xm7t"] Oct 01 16:26:10 crc kubenswrapper[4726]: I1001 16:26:10.034945 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-6bf9-account-create-9xm7t"] Oct 01 16:26:11 crc kubenswrapper[4726]: I1001 16:26:11.821515 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4" path="/var/lib/kubelet/pods/1bab7c18-2d6c-46a2-8b5a-90a0c441a0f4/volumes" Oct 01 16:26:17 crc kubenswrapper[4726]: I1001 16:26:17.814538 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:26:17 crc kubenswrapper[4726]: E1001 16:26:17.815626 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:26:20 crc kubenswrapper[4726]: I1001 16:26:20.033087 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-1b50-account-create-qw2dx"] Oct 01 16:26:20 crc kubenswrapper[4726]: I1001 16:26:20.048961 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-1b50-account-create-qw2dx"] Oct 01 16:26:21 crc kubenswrapper[4726]: I1001 16:26:21.660382 4726 scope.go:117] "RemoveContainer" containerID="00ef31a6c2ac6402d402e0304aad181a6860208742aeb4be38bb86509da08d98" Oct 01 16:26:21 crc kubenswrapper[4726]: I1001 16:26:21.682463 4726 scope.go:117] "RemoveContainer" containerID="913c7a097b09c995bef106caf5cad831590479a065d7455706a5989f0e437bba" Oct 01 16:26:21 crc kubenswrapper[4726]: I1001 16:26:21.728967 4726 scope.go:117] "RemoveContainer" containerID="606f5e42642c8a16c99a89877d19df62fa1d606137bd5c7bd896ad74f51bbf27" Oct 01 16:26:21 crc kubenswrapper[4726]: I1001 16:26:21.780425 4726 scope.go:117] "RemoveContainer" containerID="a4eed3fca647d776718f3159f4644836840c3c9cee56d8b1260375375127d5e5" Oct 01 16:26:21 crc kubenswrapper[4726]: I1001 16:26:21.819385 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="708ea401-1d87-4a51-ab33-b9318fe542e2" path="/var/lib/kubelet/pods/708ea401-1d87-4a51-ab33-b9318fe542e2/volumes" Oct 01 16:26:21 crc kubenswrapper[4726]: I1001 16:26:21.821335 4726 scope.go:117] "RemoveContainer" containerID="c87bf65c6246937eb2398b3c487d03304986c354f5e8f24957cc8aa27aa5b7d8" Oct 01 16:26:21 crc kubenswrapper[4726]: I1001 16:26:21.871231 4726 scope.go:117] "RemoveContainer" containerID="48cb20845a5026054f33727a088ddf596c76343f7fbe91163c232e0c334762dc" Oct 01 16:26:21 crc kubenswrapper[4726]: I1001 16:26:21.925179 4726 scope.go:117] "RemoveContainer" containerID="a9a17ae1b6d831f93318caa0fc9fd8249b771d2b7792ffa6ff4d089140f40ee8" Oct 01 16:26:21 crc kubenswrapper[4726]: I1001 16:26:21.955471 4726 scope.go:117] "RemoveContainer" containerID="ecb0b0dac1d3a368c9250b62f6fee94ee5de5548750f72d5da165556dd8b3c5e" Oct 01 16:26:21 crc kubenswrapper[4726]: I1001 16:26:21.982541 4726 scope.go:117] "RemoveContainer" containerID="14333894a4ee2c149a52cce8efff90eba2703e7d963369d3b26cc6f56a1f2c0d" Oct 01 16:26:22 crc kubenswrapper[4726]: I1001 16:26:22.021805 4726 scope.go:117] "RemoveContainer" containerID="778ae1d844e80e8a01605ecdf5b8268efe10337575ffb4ec750b9f8d0e97c270" Oct 01 16:26:22 crc kubenswrapper[4726]: I1001 16:26:22.048032 4726 scope.go:117] "RemoveContainer" containerID="5acce15b79fcc32e2b7f5ff77c8e1e597a20080a3c4105db28970445022b8206" Oct 01 16:26:22 crc kubenswrapper[4726]: I1001 16:26:22.083447 4726 scope.go:117] "RemoveContainer" containerID="61b7397e92066dbc8c6f9e5041df59102d2d6a4422ec5a955752453831739c9e" Oct 01 16:26:29 crc kubenswrapper[4726]: I1001 16:26:29.808991 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:26:29 crc kubenswrapper[4726]: E1001 16:26:29.810902 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:26:31 crc kubenswrapper[4726]: I1001 16:26:31.052771 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-2df5f"] Oct 01 16:26:31 crc kubenswrapper[4726]: I1001 16:26:31.069536 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-2df5f"] Oct 01 16:26:31 crc kubenswrapper[4726]: I1001 16:26:31.826282 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a875fd8-855b-48b4-91ad-043ad70711a1" path="/var/lib/kubelet/pods/4a875fd8-855b-48b4-91ad-043ad70711a1/volumes" Oct 01 16:26:43 crc kubenswrapper[4726]: I1001 16:26:43.808045 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:26:43 crc kubenswrapper[4726]: E1001 16:26:43.808917 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:26:55 crc kubenswrapper[4726]: I1001 16:26:55.808368 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:26:55 crc kubenswrapper[4726]: E1001 16:26:55.809572 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:27:04 crc kubenswrapper[4726]: I1001 16:27:04.613694 4726 generic.go:334] "Generic (PLEG): container finished" podID="bf27af63-5f62-4d0b-98e6-a7a42fb80d7b" containerID="dee6640eb5b49790ea972e7db284ffb2ce6aa1488f05a77f0c45bb4aceecad32" exitCode=0 Oct 01 16:27:04 crc kubenswrapper[4726]: I1001 16:27:04.613807 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" event={"ID":"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b","Type":"ContainerDied","Data":"dee6640eb5b49790ea972e7db284ffb2ce6aa1488f05a77f0c45bb4aceecad32"} Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.035815 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.152737 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-ssh-key\") pod \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\" (UID: \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\") " Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.153043 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5rs4\" (UniqueName: \"kubernetes.io/projected/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-kube-api-access-c5rs4\") pod \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\" (UID: \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\") " Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.153108 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-inventory\") pod \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\" (UID: \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\") " Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.168377 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-kube-api-access-c5rs4" (OuterVolumeSpecName: "kube-api-access-c5rs4") pod "bf27af63-5f62-4d0b-98e6-a7a42fb80d7b" (UID: "bf27af63-5f62-4d0b-98e6-a7a42fb80d7b"). InnerVolumeSpecName "kube-api-access-c5rs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:27:06 crc kubenswrapper[4726]: E1001 16:27:06.175803 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-inventory podName:bf27af63-5f62-4d0b-98e6-a7a42fb80d7b nodeName:}" failed. No retries permitted until 2025-10-01 16:27:06.675780066 +0000 UTC m=+1679.577332643 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-inventory") pod "bf27af63-5f62-4d0b-98e6-a7a42fb80d7b" (UID: "bf27af63-5f62-4d0b-98e6-a7a42fb80d7b") : error deleting /var/lib/kubelet/pods/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b/volume-subpaths: remove /var/lib/kubelet/pods/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b/volume-subpaths: no such file or directory Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.178791 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bf27af63-5f62-4d0b-98e6-a7a42fb80d7b" (UID: "bf27af63-5f62-4d0b-98e6-a7a42fb80d7b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.255969 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5rs4\" (UniqueName: \"kubernetes.io/projected/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-kube-api-access-c5rs4\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.256008 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.636525 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" event={"ID":"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b","Type":"ContainerDied","Data":"61008bdb23060b1b3876491ac32503768685bb89710f988a531694567cbfa49d"} Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.636578 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61008bdb23060b1b3876491ac32503768685bb89710f988a531694567cbfa49d" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.636672 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-g7c58" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.711810 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz"] Oct 01 16:27:06 crc kubenswrapper[4726]: E1001 16:27:06.712267 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="299aa258-7b3f-4759-bfc2-abeecf066d81" containerName="registry-server" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.712289 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="299aa258-7b3f-4759-bfc2-abeecf066d81" containerName="registry-server" Oct 01 16:27:06 crc kubenswrapper[4726]: E1001 16:27:06.712315 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf27af63-5f62-4d0b-98e6-a7a42fb80d7b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.712324 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf27af63-5f62-4d0b-98e6-a7a42fb80d7b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 01 16:27:06 crc kubenswrapper[4726]: E1001 16:27:06.712351 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="299aa258-7b3f-4759-bfc2-abeecf066d81" containerName="extract-utilities" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.712360 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="299aa258-7b3f-4759-bfc2-abeecf066d81" containerName="extract-utilities" Oct 01 16:27:06 crc kubenswrapper[4726]: E1001 16:27:06.712386 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="299aa258-7b3f-4759-bfc2-abeecf066d81" containerName="extract-content" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.712394 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="299aa258-7b3f-4759-bfc2-abeecf066d81" containerName="extract-content" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.712642 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="299aa258-7b3f-4759-bfc2-abeecf066d81" containerName="registry-server" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.712662 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf27af63-5f62-4d0b-98e6-a7a42fb80d7b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.713322 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.721271 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz"] Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.768012 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-inventory\") pod \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\" (UID: \"bf27af63-5f62-4d0b-98e6-a7a42fb80d7b\") " Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.779776 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-inventory" (OuterVolumeSpecName: "inventory") pod "bf27af63-5f62-4d0b-98e6-a7a42fb80d7b" (UID: "bf27af63-5f62-4d0b-98e6-a7a42fb80d7b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.876327 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec513d51-0278-4d26-aaa7-a66264c32356-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fttzz\" (UID: \"ec513d51-0278-4d26-aaa7-a66264c32356\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.876412 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec513d51-0278-4d26-aaa7-a66264c32356-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fttzz\" (UID: \"ec513d51-0278-4d26-aaa7-a66264c32356\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.876470 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvw2r\" (UniqueName: \"kubernetes.io/projected/ec513d51-0278-4d26-aaa7-a66264c32356-kube-api-access-fvw2r\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fttzz\" (UID: \"ec513d51-0278-4d26-aaa7-a66264c32356\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.876526 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf27af63-5f62-4d0b-98e6-a7a42fb80d7b-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.977606 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec513d51-0278-4d26-aaa7-a66264c32356-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fttzz\" (UID: \"ec513d51-0278-4d26-aaa7-a66264c32356\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.977687 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec513d51-0278-4d26-aaa7-a66264c32356-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fttzz\" (UID: \"ec513d51-0278-4d26-aaa7-a66264c32356\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.977784 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvw2r\" (UniqueName: \"kubernetes.io/projected/ec513d51-0278-4d26-aaa7-a66264c32356-kube-api-access-fvw2r\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fttzz\" (UID: \"ec513d51-0278-4d26-aaa7-a66264c32356\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.981742 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec513d51-0278-4d26-aaa7-a66264c32356-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fttzz\" (UID: \"ec513d51-0278-4d26-aaa7-a66264c32356\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.982456 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec513d51-0278-4d26-aaa7-a66264c32356-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fttzz\" (UID: \"ec513d51-0278-4d26-aaa7-a66264c32356\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" Oct 01 16:27:06 crc kubenswrapper[4726]: I1001 16:27:06.999875 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvw2r\" (UniqueName: \"kubernetes.io/projected/ec513d51-0278-4d26-aaa7-a66264c32356-kube-api-access-fvw2r\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fttzz\" (UID: \"ec513d51-0278-4d26-aaa7-a66264c32356\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" Oct 01 16:27:07 crc kubenswrapper[4726]: I1001 16:27:07.032126 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" Oct 01 16:27:07 crc kubenswrapper[4726]: I1001 16:27:07.553033 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz"] Oct 01 16:27:07 crc kubenswrapper[4726]: W1001 16:27:07.559880 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec513d51_0278_4d26_aaa7_a66264c32356.slice/crio-51b18049dc58f1ff385069ad88a61eb027bac538c1b2e381b9d0ae4d4fb9152f WatchSource:0}: Error finding container 51b18049dc58f1ff385069ad88a61eb027bac538c1b2e381b9d0ae4d4fb9152f: Status 404 returned error can't find the container with id 51b18049dc58f1ff385069ad88a61eb027bac538c1b2e381b9d0ae4d4fb9152f Oct 01 16:27:07 crc kubenswrapper[4726]: I1001 16:27:07.646529 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" event={"ID":"ec513d51-0278-4d26-aaa7-a66264c32356","Type":"ContainerStarted","Data":"51b18049dc58f1ff385069ad88a61eb027bac538c1b2e381b9d0ae4d4fb9152f"} Oct 01 16:27:08 crc kubenswrapper[4726]: I1001 16:27:08.808744 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:27:08 crc kubenswrapper[4726]: E1001 16:27:08.809382 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:27:09 crc kubenswrapper[4726]: I1001 16:27:09.665523 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" event={"ID":"ec513d51-0278-4d26-aaa7-a66264c32356","Type":"ContainerStarted","Data":"6c89f23dae980745ce08921c69d7f97649fb7d9cfa256ab793b6554279e51822"} Oct 01 16:27:09 crc kubenswrapper[4726]: I1001 16:27:09.694371 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" podStartSLOduration=2.302373304 podStartE2EDuration="3.694350229s" podCreationTimestamp="2025-10-01 16:27:06 +0000 UTC" firstStartedPulling="2025-10-01 16:27:07.562814059 +0000 UTC m=+1680.464366636" lastFinishedPulling="2025-10-01 16:27:08.954790984 +0000 UTC m=+1681.856343561" observedRunningTime="2025-10-01 16:27:09.678982494 +0000 UTC m=+1682.580535071" watchObservedRunningTime="2025-10-01 16:27:09.694350229 +0000 UTC m=+1682.595902806" Oct 01 16:27:15 crc kubenswrapper[4726]: I1001 16:27:15.041173 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-zwfrl"] Oct 01 16:27:15 crc kubenswrapper[4726]: I1001 16:27:15.050158 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-zwfrl"] Oct 01 16:27:15 crc kubenswrapper[4726]: I1001 16:27:15.822364 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdf2f47d-bf60-4b49-ae78-1b7823838797" path="/var/lib/kubelet/pods/bdf2f47d-bf60-4b49-ae78-1b7823838797/volumes" Oct 01 16:27:22 crc kubenswrapper[4726]: I1001 16:27:22.472587 4726 scope.go:117] "RemoveContainer" containerID="5d95995d161ada2dffe5ebe08d7905b3a6343395a2fa4d121ffbfa5dd587f3c5" Oct 01 16:27:22 crc kubenswrapper[4726]: I1001 16:27:22.512862 4726 scope.go:117] "RemoveContainer" containerID="49212a1bbb2d4fb57572e6e200ff3a2d957958def82921f59307fad9a1fde616" Oct 01 16:27:23 crc kubenswrapper[4726]: I1001 16:27:23.808637 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:27:23 crc kubenswrapper[4726]: E1001 16:27:23.809400 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:27:27 crc kubenswrapper[4726]: I1001 16:27:27.040345 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-zfnx8"] Oct 01 16:27:27 crc kubenswrapper[4726]: I1001 16:27:27.047468 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-zfnx8"] Oct 01 16:27:27 crc kubenswrapper[4726]: I1001 16:27:27.827767 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df1f486b-c5af-47de-9af6-4c499af830db" path="/var/lib/kubelet/pods/df1f486b-c5af-47de-9af6-4c499af830db/volumes" Oct 01 16:27:28 crc kubenswrapper[4726]: I1001 16:27:28.031718 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-ks2wk"] Oct 01 16:27:28 crc kubenswrapper[4726]: I1001 16:27:28.040853 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-ks2wk"] Oct 01 16:27:29 crc kubenswrapper[4726]: I1001 16:27:29.820213 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c7939fb-33c3-486d-85a6-05265f29d3e8" path="/var/lib/kubelet/pods/9c7939fb-33c3-486d-85a6-05265f29d3e8/volumes" Oct 01 16:27:35 crc kubenswrapper[4726]: I1001 16:27:35.810611 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:27:35 crc kubenswrapper[4726]: E1001 16:27:35.811730 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:27:46 crc kubenswrapper[4726]: I1001 16:27:46.808423 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:27:46 crc kubenswrapper[4726]: E1001 16:27:46.809362 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:27:56 crc kubenswrapper[4726]: I1001 16:27:56.054872 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-bhxd2"] Oct 01 16:27:56 crc kubenswrapper[4726]: I1001 16:27:56.069388 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-bhxd2"] Oct 01 16:27:57 crc kubenswrapper[4726]: I1001 16:27:57.037888 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-gsljc"] Oct 01 16:27:57 crc kubenswrapper[4726]: I1001 16:27:57.048173 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-gsljc"] Oct 01 16:27:57 crc kubenswrapper[4726]: I1001 16:27:57.822266 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a83273b9-e1f0-47e6-8bbe-839f4c0075b3" path="/var/lib/kubelet/pods/a83273b9-e1f0-47e6-8bbe-839f4c0075b3/volumes" Oct 01 16:27:57 crc kubenswrapper[4726]: I1001 16:27:57.823982 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6" path="/var/lib/kubelet/pods/fdb5fc73-43ce-425a-8a5f-cd1b3b96aba6/volumes" Oct 01 16:27:58 crc kubenswrapper[4726]: I1001 16:27:58.035077 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-skmf7"] Oct 01 16:27:58 crc kubenswrapper[4726]: I1001 16:27:58.048792 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-skmf7"] Oct 01 16:27:59 crc kubenswrapper[4726]: I1001 16:27:59.820547 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b994ac8-f28c-458a-873f-4270e9f0f36b" path="/var/lib/kubelet/pods/7b994ac8-f28c-458a-873f-4270e9f0f36b/volumes" Oct 01 16:28:01 crc kubenswrapper[4726]: I1001 16:28:01.809460 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:28:01 crc kubenswrapper[4726]: E1001 16:28:01.810192 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:28:14 crc kubenswrapper[4726]: I1001 16:28:14.809662 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:28:14 crc kubenswrapper[4726]: E1001 16:28:14.810565 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:28:19 crc kubenswrapper[4726]: I1001 16:28:19.335515 4726 generic.go:334] "Generic (PLEG): container finished" podID="ec513d51-0278-4d26-aaa7-a66264c32356" containerID="6c89f23dae980745ce08921c69d7f97649fb7d9cfa256ab793b6554279e51822" exitCode=0 Oct 01 16:28:19 crc kubenswrapper[4726]: I1001 16:28:19.335601 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" event={"ID":"ec513d51-0278-4d26-aaa7-a66264c32356","Type":"ContainerDied","Data":"6c89f23dae980745ce08921c69d7f97649fb7d9cfa256ab793b6554279e51822"} Oct 01 16:28:20 crc kubenswrapper[4726]: I1001 16:28:20.728575 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" Oct 01 16:28:20 crc kubenswrapper[4726]: I1001 16:28:20.863563 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvw2r\" (UniqueName: \"kubernetes.io/projected/ec513d51-0278-4d26-aaa7-a66264c32356-kube-api-access-fvw2r\") pod \"ec513d51-0278-4d26-aaa7-a66264c32356\" (UID: \"ec513d51-0278-4d26-aaa7-a66264c32356\") " Oct 01 16:28:20 crc kubenswrapper[4726]: I1001 16:28:20.863645 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec513d51-0278-4d26-aaa7-a66264c32356-inventory\") pod \"ec513d51-0278-4d26-aaa7-a66264c32356\" (UID: \"ec513d51-0278-4d26-aaa7-a66264c32356\") " Oct 01 16:28:20 crc kubenswrapper[4726]: I1001 16:28:20.863753 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec513d51-0278-4d26-aaa7-a66264c32356-ssh-key\") pod \"ec513d51-0278-4d26-aaa7-a66264c32356\" (UID: \"ec513d51-0278-4d26-aaa7-a66264c32356\") " Oct 01 16:28:20 crc kubenswrapper[4726]: I1001 16:28:20.869570 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec513d51-0278-4d26-aaa7-a66264c32356-kube-api-access-fvw2r" (OuterVolumeSpecName: "kube-api-access-fvw2r") pod "ec513d51-0278-4d26-aaa7-a66264c32356" (UID: "ec513d51-0278-4d26-aaa7-a66264c32356"). InnerVolumeSpecName "kube-api-access-fvw2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:28:20 crc kubenswrapper[4726]: I1001 16:28:20.907553 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec513d51-0278-4d26-aaa7-a66264c32356-inventory" (OuterVolumeSpecName: "inventory") pod "ec513d51-0278-4d26-aaa7-a66264c32356" (UID: "ec513d51-0278-4d26-aaa7-a66264c32356"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:28:20 crc kubenswrapper[4726]: I1001 16:28:20.908530 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec513d51-0278-4d26-aaa7-a66264c32356-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ec513d51-0278-4d26-aaa7-a66264c32356" (UID: "ec513d51-0278-4d26-aaa7-a66264c32356"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:28:20 crc kubenswrapper[4726]: I1001 16:28:20.966847 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvw2r\" (UniqueName: \"kubernetes.io/projected/ec513d51-0278-4d26-aaa7-a66264c32356-kube-api-access-fvw2r\") on node \"crc\" DevicePath \"\"" Oct 01 16:28:20 crc kubenswrapper[4726]: I1001 16:28:20.966899 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec513d51-0278-4d26-aaa7-a66264c32356-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:28:20 crc kubenswrapper[4726]: I1001 16:28:20.966917 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec513d51-0278-4d26-aaa7-a66264c32356-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.362444 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" event={"ID":"ec513d51-0278-4d26-aaa7-a66264c32356","Type":"ContainerDied","Data":"51b18049dc58f1ff385069ad88a61eb027bac538c1b2e381b9d0ae4d4fb9152f"} Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.362823 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51b18049dc58f1ff385069ad88a61eb027bac538c1b2e381b9d0ae4d4fb9152f" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.362555 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fttzz" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.494184 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk"] Oct 01 16:28:21 crc kubenswrapper[4726]: E1001 16:28:21.495278 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec513d51-0278-4d26-aaa7-a66264c32356" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.495304 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec513d51-0278-4d26-aaa7-a66264c32356" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.495851 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec513d51-0278-4d26-aaa7-a66264c32356" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.498180 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.500466 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.500848 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.501042 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.500888 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.509349 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk"] Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.583797 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk\" (UID: \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.583911 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk\" (UID: \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.583972 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss9lg\" (UniqueName: \"kubernetes.io/projected/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-kube-api-access-ss9lg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk\" (UID: \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.686021 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk\" (UID: \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.686113 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss9lg\" (UniqueName: \"kubernetes.io/projected/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-kube-api-access-ss9lg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk\" (UID: \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.686268 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk\" (UID: \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.691142 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk\" (UID: \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.695011 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk\" (UID: \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.705724 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss9lg\" (UniqueName: \"kubernetes.io/projected/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-kube-api-access-ss9lg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk\" (UID: \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" Oct 01 16:28:21 crc kubenswrapper[4726]: I1001 16:28:21.822492 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" Oct 01 16:28:22 crc kubenswrapper[4726]: I1001 16:28:22.339557 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk"] Oct 01 16:28:22 crc kubenswrapper[4726]: W1001 16:28:22.351800 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc69a5dd3_367d_4db8_907c_cb7afb6da6a9.slice/crio-a9bccd3349a85b2e8c2ed5cbe44e42098bf7722158cbc7efedc736142f6e8066 WatchSource:0}: Error finding container a9bccd3349a85b2e8c2ed5cbe44e42098bf7722158cbc7efedc736142f6e8066: Status 404 returned error can't find the container with id a9bccd3349a85b2e8c2ed5cbe44e42098bf7722158cbc7efedc736142f6e8066 Oct 01 16:28:22 crc kubenswrapper[4726]: I1001 16:28:22.372493 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" event={"ID":"c69a5dd3-367d-4db8-907c-cb7afb6da6a9","Type":"ContainerStarted","Data":"a9bccd3349a85b2e8c2ed5cbe44e42098bf7722158cbc7efedc736142f6e8066"} Oct 01 16:28:22 crc kubenswrapper[4726]: I1001 16:28:22.602952 4726 scope.go:117] "RemoveContainer" containerID="0fd712d7373def2d15480febabcafaa3ba5119977c938919054b21fd5ef60748" Oct 01 16:28:22 crc kubenswrapper[4726]: I1001 16:28:22.630239 4726 scope.go:117] "RemoveContainer" containerID="09be41a89fb7158684759b4ae7e3644120e39d857595421392f69ecee43fd0d4" Oct 01 16:28:22 crc kubenswrapper[4726]: I1001 16:28:22.663591 4726 scope.go:117] "RemoveContainer" containerID="16d3ded60afe96d427098a307e71f20d4f35f12993be114dad1f6c2582f663a1" Oct 01 16:28:22 crc kubenswrapper[4726]: I1001 16:28:22.714221 4726 scope.go:117] "RemoveContainer" containerID="ea492a4fbf49900bab734a0b934d107aece9b120a65c85bd355aff4831b1d322" Oct 01 16:28:22 crc kubenswrapper[4726]: I1001 16:28:22.767932 4726 scope.go:117] "RemoveContainer" containerID="79416f7542f476fdc76e78a15e310f7390f6670c401eb02b3b186018905708d4" Oct 01 16:28:23 crc kubenswrapper[4726]: I1001 16:28:23.381449 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" event={"ID":"c69a5dd3-367d-4db8-907c-cb7afb6da6a9","Type":"ContainerStarted","Data":"a5465b600d17d8d04e21ce12512473629486397dcd926fc228a9b9e67bf32bfc"} Oct 01 16:28:23 crc kubenswrapper[4726]: I1001 16:28:23.399305 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" podStartSLOduration=1.799676948 podStartE2EDuration="2.399285629s" podCreationTimestamp="2025-10-01 16:28:21 +0000 UTC" firstStartedPulling="2025-10-01 16:28:22.354812316 +0000 UTC m=+1755.256364913" lastFinishedPulling="2025-10-01 16:28:22.954421017 +0000 UTC m=+1755.855973594" observedRunningTime="2025-10-01 16:28:23.394621253 +0000 UTC m=+1756.296173850" watchObservedRunningTime="2025-10-01 16:28:23.399285629 +0000 UTC m=+1756.300838206" Oct 01 16:28:28 crc kubenswrapper[4726]: I1001 16:28:28.437123 4726 generic.go:334] "Generic (PLEG): container finished" podID="c69a5dd3-367d-4db8-907c-cb7afb6da6a9" containerID="a5465b600d17d8d04e21ce12512473629486397dcd926fc228a9b9e67bf32bfc" exitCode=0 Oct 01 16:28:28 crc kubenswrapper[4726]: I1001 16:28:28.437216 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" event={"ID":"c69a5dd3-367d-4db8-907c-cb7afb6da6a9","Type":"ContainerDied","Data":"a5465b600d17d8d04e21ce12512473629486397dcd926fc228a9b9e67bf32bfc"} Oct 01 16:28:28 crc kubenswrapper[4726]: I1001 16:28:28.807866 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:28:28 crc kubenswrapper[4726]: E1001 16:28:28.808522 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:28:29 crc kubenswrapper[4726]: I1001 16:28:29.850007 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" Oct 01 16:28:29 crc kubenswrapper[4726]: I1001 16:28:29.942898 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss9lg\" (UniqueName: \"kubernetes.io/projected/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-kube-api-access-ss9lg\") pod \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\" (UID: \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\") " Oct 01 16:28:29 crc kubenswrapper[4726]: I1001 16:28:29.943183 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-inventory\") pod \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\" (UID: \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\") " Oct 01 16:28:29 crc kubenswrapper[4726]: I1001 16:28:29.943335 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-ssh-key\") pod \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\" (UID: \"c69a5dd3-367d-4db8-907c-cb7afb6da6a9\") " Oct 01 16:28:29 crc kubenswrapper[4726]: I1001 16:28:29.949611 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-kube-api-access-ss9lg" (OuterVolumeSpecName: "kube-api-access-ss9lg") pod "c69a5dd3-367d-4db8-907c-cb7afb6da6a9" (UID: "c69a5dd3-367d-4db8-907c-cb7afb6da6a9"). InnerVolumeSpecName "kube-api-access-ss9lg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:28:29 crc kubenswrapper[4726]: I1001 16:28:29.972253 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-inventory" (OuterVolumeSpecName: "inventory") pod "c69a5dd3-367d-4db8-907c-cb7afb6da6a9" (UID: "c69a5dd3-367d-4db8-907c-cb7afb6da6a9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:28:29 crc kubenswrapper[4726]: I1001 16:28:29.993788 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c69a5dd3-367d-4db8-907c-cb7afb6da6a9" (UID: "c69a5dd3-367d-4db8-907c-cb7afb6da6a9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.044114 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-hvk5c"] Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.045837 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.045863 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.045875 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss9lg\" (UniqueName: \"kubernetes.io/projected/c69a5dd3-367d-4db8-907c-cb7afb6da6a9-kube-api-access-ss9lg\") on node \"crc\" DevicePath \"\"" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.057115 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-8v9wq"] Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.063702 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-kcw4f"] Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.070746 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-hvk5c"] Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.077449 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-8v9wq"] Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.083625 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-kcw4f"] Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.458829 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" event={"ID":"c69a5dd3-367d-4db8-907c-cb7afb6da6a9","Type":"ContainerDied","Data":"a9bccd3349a85b2e8c2ed5cbe44e42098bf7722158cbc7efedc736142f6e8066"} Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.458872 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9bccd3349a85b2e8c2ed5cbe44e42098bf7722158cbc7efedc736142f6e8066" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.458986 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.562604 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2"] Oct 01 16:28:30 crc kubenswrapper[4726]: E1001 16:28:30.562991 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c69a5dd3-367d-4db8-907c-cb7afb6da6a9" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.563009 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c69a5dd3-367d-4db8-907c-cb7afb6da6a9" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.563217 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c69a5dd3-367d-4db8-907c-cb7afb6da6a9" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.563862 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.567479 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.568577 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.569508 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.574420 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.583006 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2"] Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.654823 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4727b\" (UniqueName: \"kubernetes.io/projected/4cdcd93d-244a-4adb-b295-ae609306b749-kube-api-access-4727b\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qrzm2\" (UID: \"4cdcd93d-244a-4adb-b295-ae609306b749\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.654995 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4cdcd93d-244a-4adb-b295-ae609306b749-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qrzm2\" (UID: \"4cdcd93d-244a-4adb-b295-ae609306b749\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.655036 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4cdcd93d-244a-4adb-b295-ae609306b749-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qrzm2\" (UID: \"4cdcd93d-244a-4adb-b295-ae609306b749\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.756965 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4cdcd93d-244a-4adb-b295-ae609306b749-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qrzm2\" (UID: \"4cdcd93d-244a-4adb-b295-ae609306b749\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.757029 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4cdcd93d-244a-4adb-b295-ae609306b749-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qrzm2\" (UID: \"4cdcd93d-244a-4adb-b295-ae609306b749\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.757115 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4727b\" (UniqueName: \"kubernetes.io/projected/4cdcd93d-244a-4adb-b295-ae609306b749-kube-api-access-4727b\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qrzm2\" (UID: \"4cdcd93d-244a-4adb-b295-ae609306b749\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.761417 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4cdcd93d-244a-4adb-b295-ae609306b749-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qrzm2\" (UID: \"4cdcd93d-244a-4adb-b295-ae609306b749\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.761689 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4cdcd93d-244a-4adb-b295-ae609306b749-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qrzm2\" (UID: \"4cdcd93d-244a-4adb-b295-ae609306b749\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.772975 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4727b\" (UniqueName: \"kubernetes.io/projected/4cdcd93d-244a-4adb-b295-ae609306b749-kube-api-access-4727b\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qrzm2\" (UID: \"4cdcd93d-244a-4adb-b295-ae609306b749\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" Oct 01 16:28:30 crc kubenswrapper[4726]: I1001 16:28:30.885812 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" Oct 01 16:28:31 crc kubenswrapper[4726]: I1001 16:28:31.355816 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2"] Oct 01 16:28:31 crc kubenswrapper[4726]: I1001 16:28:31.470283 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" event={"ID":"4cdcd93d-244a-4adb-b295-ae609306b749","Type":"ContainerStarted","Data":"48d29df6619f45d59b1564ac72fa04e087e0b1114ff47d372f1f072c82bf1e1c"} Oct 01 16:28:31 crc kubenswrapper[4726]: I1001 16:28:31.833774 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ff89aa8-5b64-461c-9cad-7fe1e0375c77" path="/var/lib/kubelet/pods/1ff89aa8-5b64-461c-9cad-7fe1e0375c77/volumes" Oct 01 16:28:31 crc kubenswrapper[4726]: I1001 16:28:31.835806 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d0b9d80-050e-47bf-a71d-1ca4b10ac868" path="/var/lib/kubelet/pods/9d0b9d80-050e-47bf-a71d-1ca4b10ac868/volumes" Oct 01 16:28:31 crc kubenswrapper[4726]: I1001 16:28:31.836517 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5eadc81-68d4-44be-acd8-6127518dcbdf" path="/var/lib/kubelet/pods/c5eadc81-68d4-44be-acd8-6127518dcbdf/volumes" Oct 01 16:28:32 crc kubenswrapper[4726]: I1001 16:28:32.481982 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" event={"ID":"4cdcd93d-244a-4adb-b295-ae609306b749","Type":"ContainerStarted","Data":"737af3a26739fa3de27c5a7e521231a1b0bed80c2bcb5e44d2563829f1d98095"} Oct 01 16:28:32 crc kubenswrapper[4726]: I1001 16:28:32.507278 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" podStartSLOduration=1.842179293 podStartE2EDuration="2.507259833s" podCreationTimestamp="2025-10-01 16:28:30 +0000 UTC" firstStartedPulling="2025-10-01 16:28:31.361684187 +0000 UTC m=+1764.263236764" lastFinishedPulling="2025-10-01 16:28:32.026764727 +0000 UTC m=+1764.928317304" observedRunningTime="2025-10-01 16:28:32.502870806 +0000 UTC m=+1765.404423393" watchObservedRunningTime="2025-10-01 16:28:32.507259833 +0000 UTC m=+1765.408812410" Oct 01 16:28:40 crc kubenswrapper[4726]: I1001 16:28:40.034455 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-f3eb-account-create-6g7cm"] Oct 01 16:28:40 crc kubenswrapper[4726]: I1001 16:28:40.047943 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-560d-account-create-fztkr"] Oct 01 16:28:40 crc kubenswrapper[4726]: I1001 16:28:40.074375 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-b942-account-create-7x8nm"] Oct 01 16:28:40 crc kubenswrapper[4726]: I1001 16:28:40.084449 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-560d-account-create-fztkr"] Oct 01 16:28:40 crc kubenswrapper[4726]: I1001 16:28:40.092355 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-f3eb-account-create-6g7cm"] Oct 01 16:28:40 crc kubenswrapper[4726]: I1001 16:28:40.098816 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-b942-account-create-7x8nm"] Oct 01 16:28:41 crc kubenswrapper[4726]: I1001 16:28:41.820587 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c7c4328-6015-4e9a-9cd8-67c1d275928e" path="/var/lib/kubelet/pods/1c7c4328-6015-4e9a-9cd8-67c1d275928e/volumes" Oct 01 16:28:41 crc kubenswrapper[4726]: I1001 16:28:41.821707 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82d29008-5715-4a39-b0ea-252402970cdc" path="/var/lib/kubelet/pods/82d29008-5715-4a39-b0ea-252402970cdc/volumes" Oct 01 16:28:41 crc kubenswrapper[4726]: I1001 16:28:41.822347 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7a05e67-296b-4ed9-b63c-30c6edc152db" path="/var/lib/kubelet/pods/a7a05e67-296b-4ed9-b63c-30c6edc152db/volumes" Oct 01 16:28:43 crc kubenswrapper[4726]: I1001 16:28:43.808614 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:28:43 crc kubenswrapper[4726]: E1001 16:28:43.809463 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:28:54 crc kubenswrapper[4726]: I1001 16:28:54.808770 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:28:54 crc kubenswrapper[4726]: E1001 16:28:54.809627 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:29:07 crc kubenswrapper[4726]: I1001 16:29:07.832809 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:29:07 crc kubenswrapper[4726]: E1001 16:29:07.833520 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:29:10 crc kubenswrapper[4726]: I1001 16:29:10.855702 4726 generic.go:334] "Generic (PLEG): container finished" podID="4cdcd93d-244a-4adb-b295-ae609306b749" containerID="737af3a26739fa3de27c5a7e521231a1b0bed80c2bcb5e44d2563829f1d98095" exitCode=0 Oct 01 16:29:10 crc kubenswrapper[4726]: I1001 16:29:10.855787 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" event={"ID":"4cdcd93d-244a-4adb-b295-ae609306b749","Type":"ContainerDied","Data":"737af3a26739fa3de27c5a7e521231a1b0bed80c2bcb5e44d2563829f1d98095"} Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.358749 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.405558 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4cdcd93d-244a-4adb-b295-ae609306b749-ssh-key\") pod \"4cdcd93d-244a-4adb-b295-ae609306b749\" (UID: \"4cdcd93d-244a-4adb-b295-ae609306b749\") " Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.405736 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4cdcd93d-244a-4adb-b295-ae609306b749-inventory\") pod \"4cdcd93d-244a-4adb-b295-ae609306b749\" (UID: \"4cdcd93d-244a-4adb-b295-ae609306b749\") " Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.405807 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4727b\" (UniqueName: \"kubernetes.io/projected/4cdcd93d-244a-4adb-b295-ae609306b749-kube-api-access-4727b\") pod \"4cdcd93d-244a-4adb-b295-ae609306b749\" (UID: \"4cdcd93d-244a-4adb-b295-ae609306b749\") " Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.412401 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cdcd93d-244a-4adb-b295-ae609306b749-kube-api-access-4727b" (OuterVolumeSpecName: "kube-api-access-4727b") pod "4cdcd93d-244a-4adb-b295-ae609306b749" (UID: "4cdcd93d-244a-4adb-b295-ae609306b749"). InnerVolumeSpecName "kube-api-access-4727b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.442510 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cdcd93d-244a-4adb-b295-ae609306b749-inventory" (OuterVolumeSpecName: "inventory") pod "4cdcd93d-244a-4adb-b295-ae609306b749" (UID: "4cdcd93d-244a-4adb-b295-ae609306b749"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.444886 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cdcd93d-244a-4adb-b295-ae609306b749-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4cdcd93d-244a-4adb-b295-ae609306b749" (UID: "4cdcd93d-244a-4adb-b295-ae609306b749"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.507666 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4cdcd93d-244a-4adb-b295-ae609306b749-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.507699 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4727b\" (UniqueName: \"kubernetes.io/projected/4cdcd93d-244a-4adb-b295-ae609306b749-kube-api-access-4727b\") on node \"crc\" DevicePath \"\"" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.507713 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4cdcd93d-244a-4adb-b295-ae609306b749-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.877119 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" event={"ID":"4cdcd93d-244a-4adb-b295-ae609306b749","Type":"ContainerDied","Data":"48d29df6619f45d59b1564ac72fa04e087e0b1114ff47d372f1f072c82bf1e1c"} Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.877828 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48d29df6619f45d59b1564ac72fa04e087e0b1114ff47d372f1f072c82bf1e1c" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.877267 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qrzm2" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.987110 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk"] Oct 01 16:29:12 crc kubenswrapper[4726]: E1001 16:29:12.987586 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cdcd93d-244a-4adb-b295-ae609306b749" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.987612 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cdcd93d-244a-4adb-b295-ae609306b749" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.987830 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cdcd93d-244a-4adb-b295-ae609306b749" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.988634 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.997608 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.997657 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.997858 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:29:12 crc kubenswrapper[4726]: I1001 16:29:12.998408 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:29:13 crc kubenswrapper[4726]: I1001 16:29:13.016955 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk"] Oct 01 16:29:13 crc kubenswrapper[4726]: E1001 16:29:13.034249 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4cdcd93d_244a_4adb_b295_ae609306b749.slice/crio-48d29df6619f45d59b1564ac72fa04e087e0b1114ff47d372f1f072c82bf1e1c\": RecentStats: unable to find data in memory cache]" Oct 01 16:29:13 crc kubenswrapper[4726]: I1001 16:29:13.125305 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk\" (UID: \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" Oct 01 16:29:13 crc kubenswrapper[4726]: I1001 16:29:13.125641 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk\" (UID: \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" Oct 01 16:29:13 crc kubenswrapper[4726]: I1001 16:29:13.125939 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm9x7\" (UniqueName: \"kubernetes.io/projected/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-kube-api-access-qm9x7\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk\" (UID: \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" Oct 01 16:29:13 crc kubenswrapper[4726]: I1001 16:29:13.227921 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk\" (UID: \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" Oct 01 16:29:13 crc kubenswrapper[4726]: I1001 16:29:13.228075 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk\" (UID: \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" Oct 01 16:29:13 crc kubenswrapper[4726]: I1001 16:29:13.228357 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm9x7\" (UniqueName: \"kubernetes.io/projected/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-kube-api-access-qm9x7\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk\" (UID: \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" Oct 01 16:29:13 crc kubenswrapper[4726]: I1001 16:29:13.234913 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk\" (UID: \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" Oct 01 16:29:13 crc kubenswrapper[4726]: I1001 16:29:13.241620 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk\" (UID: \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" Oct 01 16:29:13 crc kubenswrapper[4726]: I1001 16:29:13.246714 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm9x7\" (UniqueName: \"kubernetes.io/projected/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-kube-api-access-qm9x7\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk\" (UID: \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" Oct 01 16:29:13 crc kubenswrapper[4726]: I1001 16:29:13.317808 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" Oct 01 16:29:13 crc kubenswrapper[4726]: I1001 16:29:13.889570 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk"] Oct 01 16:29:13 crc kubenswrapper[4726]: W1001 16:29:13.890883 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98f0cdbd_7ea0_486b_92bc_14e78ab54ffd.slice/crio-c353e0b076595c221df764f247c170318f0f330af2299a2679214358d8a4f559 WatchSource:0}: Error finding container c353e0b076595c221df764f247c170318f0f330af2299a2679214358d8a4f559: Status 404 returned error can't find the container with id c353e0b076595c221df764f247c170318f0f330af2299a2679214358d8a4f559 Oct 01 16:29:14 crc kubenswrapper[4726]: I1001 16:29:14.896796 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" event={"ID":"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd","Type":"ContainerStarted","Data":"0b1573d7011f4191feb6f9da54a5f346a031e71958691fa4e5304f1a6970c4eb"} Oct 01 16:29:14 crc kubenswrapper[4726]: I1001 16:29:14.896862 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" event={"ID":"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd","Type":"ContainerStarted","Data":"c353e0b076595c221df764f247c170318f0f330af2299a2679214358d8a4f559"} Oct 01 16:29:14 crc kubenswrapper[4726]: I1001 16:29:14.926267 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" podStartSLOduration=2.474805845 podStartE2EDuration="2.926238168s" podCreationTimestamp="2025-10-01 16:29:12 +0000 UTC" firstStartedPulling="2025-10-01 16:29:13.894036309 +0000 UTC m=+1806.795588886" lastFinishedPulling="2025-10-01 16:29:14.345468602 +0000 UTC m=+1807.247021209" observedRunningTime="2025-10-01 16:29:14.919839592 +0000 UTC m=+1807.821392199" watchObservedRunningTime="2025-10-01 16:29:14.926238168 +0000 UTC m=+1807.827790785" Oct 01 16:29:19 crc kubenswrapper[4726]: I1001 16:29:19.045518 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hq8t6"] Oct 01 16:29:19 crc kubenswrapper[4726]: I1001 16:29:19.070846 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hq8t6"] Oct 01 16:29:19 crc kubenswrapper[4726]: I1001 16:29:19.809507 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:29:19 crc kubenswrapper[4726]: E1001 16:29:19.810190 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:29:19 crc kubenswrapper[4726]: I1001 16:29:19.822701 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="551173b5-0f09-4b1e-831b-c6598ba904cb" path="/var/lib/kubelet/pods/551173b5-0f09-4b1e-831b-c6598ba904cb/volumes" Oct 01 16:29:22 crc kubenswrapper[4726]: I1001 16:29:22.930681 4726 scope.go:117] "RemoveContainer" containerID="de70527985aba717e6a1d83443850b0ccb7a8b9147bbc29c1458868ccebc1135" Oct 01 16:29:22 crc kubenswrapper[4726]: I1001 16:29:22.965088 4726 scope.go:117] "RemoveContainer" containerID="7dbe378442bb110761661bc29105a34b7e542487981a30e0ee6ff6c42928e6cd" Oct 01 16:29:23 crc kubenswrapper[4726]: I1001 16:29:23.023530 4726 scope.go:117] "RemoveContainer" containerID="2a92e8f1fe7f8f0e9de69da93cf026e59dcdbe7e99806622f897846e76c43954" Oct 01 16:29:23 crc kubenswrapper[4726]: I1001 16:29:23.046902 4726 scope.go:117] "RemoveContainer" containerID="c501b4ff046e6ca2c04e937e150b7e1489a3ddce4f4c61c941e381e3f4f52b6a" Oct 01 16:29:23 crc kubenswrapper[4726]: I1001 16:29:23.082604 4726 scope.go:117] "RemoveContainer" containerID="5487ced454167d5e1713318c6c6a40500aa65b70bce9aade694640a0a5973e99" Oct 01 16:29:23 crc kubenswrapper[4726]: I1001 16:29:23.125235 4726 scope.go:117] "RemoveContainer" containerID="b5c623227e0851c2a7583d3c6526727451c9485f8860571c588dfc8864724dc6" Oct 01 16:29:23 crc kubenswrapper[4726]: I1001 16:29:23.163129 4726 scope.go:117] "RemoveContainer" containerID="b329989b95b7dfdbec6f920c028e2c191fcaedf3425895056eaf276a1e2f3c2a" Oct 01 16:29:32 crc kubenswrapper[4726]: I1001 16:29:32.807911 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:29:32 crc kubenswrapper[4726]: E1001 16:29:32.810542 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:29:41 crc kubenswrapper[4726]: I1001 16:29:41.046254 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-qggxs"] Oct 01 16:29:41 crc kubenswrapper[4726]: I1001 16:29:41.055948 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-qggxs"] Oct 01 16:29:41 crc kubenswrapper[4726]: I1001 16:29:41.824749 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ee77a28-a1b7-4d80-9204-42a56d4dde89" path="/var/lib/kubelet/pods/9ee77a28-a1b7-4d80-9204-42a56d4dde89/volumes" Oct 01 16:29:44 crc kubenswrapper[4726]: I1001 16:29:44.808005 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:29:44 crc kubenswrapper[4726]: E1001 16:29:44.809203 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:29:45 crc kubenswrapper[4726]: I1001 16:29:45.035519 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-79vp9"] Oct 01 16:29:45 crc kubenswrapper[4726]: I1001 16:29:45.045379 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-79vp9"] Oct 01 16:29:45 crc kubenswrapper[4726]: I1001 16:29:45.829330 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d747a1f6-2d98-4b9e-9a0d-833d0afede9d" path="/var/lib/kubelet/pods/d747a1f6-2d98-4b9e-9a0d-833d0afede9d/volumes" Oct 01 16:29:56 crc kubenswrapper[4726]: I1001 16:29:56.808638 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:29:56 crc kubenswrapper[4726]: E1001 16:29:56.810182 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:29:59 crc kubenswrapper[4726]: I1001 16:29:59.521725 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nj4wf"] Oct 01 16:29:59 crc kubenswrapper[4726]: I1001 16:29:59.527370 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:29:59 crc kubenswrapper[4726]: I1001 16:29:59.536928 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nj4wf"] Oct 01 16:29:59 crc kubenswrapper[4726]: I1001 16:29:59.644428 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c24a10a9-6d60-43d5-9e3c-67cec3e0045c-utilities\") pod \"redhat-operators-nj4wf\" (UID: \"c24a10a9-6d60-43d5-9e3c-67cec3e0045c\") " pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:29:59 crc kubenswrapper[4726]: I1001 16:29:59.644673 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c24a10a9-6d60-43d5-9e3c-67cec3e0045c-catalog-content\") pod \"redhat-operators-nj4wf\" (UID: \"c24a10a9-6d60-43d5-9e3c-67cec3e0045c\") " pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:29:59 crc kubenswrapper[4726]: I1001 16:29:59.645078 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cjbr\" (UniqueName: \"kubernetes.io/projected/c24a10a9-6d60-43d5-9e3c-67cec3e0045c-kube-api-access-6cjbr\") pod \"redhat-operators-nj4wf\" (UID: \"c24a10a9-6d60-43d5-9e3c-67cec3e0045c\") " pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:29:59 crc kubenswrapper[4726]: I1001 16:29:59.746585 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c24a10a9-6d60-43d5-9e3c-67cec3e0045c-utilities\") pod \"redhat-operators-nj4wf\" (UID: \"c24a10a9-6d60-43d5-9e3c-67cec3e0045c\") " pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:29:59 crc kubenswrapper[4726]: I1001 16:29:59.746867 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c24a10a9-6d60-43d5-9e3c-67cec3e0045c-catalog-content\") pod \"redhat-operators-nj4wf\" (UID: \"c24a10a9-6d60-43d5-9e3c-67cec3e0045c\") " pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:29:59 crc kubenswrapper[4726]: I1001 16:29:59.747001 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cjbr\" (UniqueName: \"kubernetes.io/projected/c24a10a9-6d60-43d5-9e3c-67cec3e0045c-kube-api-access-6cjbr\") pod \"redhat-operators-nj4wf\" (UID: \"c24a10a9-6d60-43d5-9e3c-67cec3e0045c\") " pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:29:59 crc kubenswrapper[4726]: I1001 16:29:59.747072 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c24a10a9-6d60-43d5-9e3c-67cec3e0045c-utilities\") pod \"redhat-operators-nj4wf\" (UID: \"c24a10a9-6d60-43d5-9e3c-67cec3e0045c\") " pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:29:59 crc kubenswrapper[4726]: I1001 16:29:59.747267 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c24a10a9-6d60-43d5-9e3c-67cec3e0045c-catalog-content\") pod \"redhat-operators-nj4wf\" (UID: \"c24a10a9-6d60-43d5-9e3c-67cec3e0045c\") " pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:29:59 crc kubenswrapper[4726]: I1001 16:29:59.769069 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cjbr\" (UniqueName: \"kubernetes.io/projected/c24a10a9-6d60-43d5-9e3c-67cec3e0045c-kube-api-access-6cjbr\") pod \"redhat-operators-nj4wf\" (UID: \"c24a10a9-6d60-43d5-9e3c-67cec3e0045c\") " pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:29:59 crc kubenswrapper[4726]: I1001 16:29:59.869005 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.161608 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd"] Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.164743 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.168950 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.169211 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.177287 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nj4wf"] Oct 01 16:30:00 crc kubenswrapper[4726]: W1001 16:30:00.181144 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc24a10a9_6d60_43d5_9e3c_67cec3e0045c.slice/crio-17ed9abfe4cc9ac1804af4d113eec15ef1aef69a292397dfb0b52a93490c8390 WatchSource:0}: Error finding container 17ed9abfe4cc9ac1804af4d113eec15ef1aef69a292397dfb0b52a93490c8390: Status 404 returned error can't find the container with id 17ed9abfe4cc9ac1804af4d113eec15ef1aef69a292397dfb0b52a93490c8390 Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.183512 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd"] Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.259152 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f684623e-fdca-4fa3-9b78-ef67b56cd76c-config-volume\") pod \"collect-profiles-29322270-qb9dd\" (UID: \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.259293 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8wg6\" (UniqueName: \"kubernetes.io/projected/f684623e-fdca-4fa3-9b78-ef67b56cd76c-kube-api-access-k8wg6\") pod \"collect-profiles-29322270-qb9dd\" (UID: \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.259334 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f684623e-fdca-4fa3-9b78-ef67b56cd76c-secret-volume\") pod \"collect-profiles-29322270-qb9dd\" (UID: \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.328587 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nj4wf" event={"ID":"c24a10a9-6d60-43d5-9e3c-67cec3e0045c","Type":"ContainerStarted","Data":"17ed9abfe4cc9ac1804af4d113eec15ef1aef69a292397dfb0b52a93490c8390"} Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.361414 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f684623e-fdca-4fa3-9b78-ef67b56cd76c-config-volume\") pod \"collect-profiles-29322270-qb9dd\" (UID: \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.361563 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8wg6\" (UniqueName: \"kubernetes.io/projected/f684623e-fdca-4fa3-9b78-ef67b56cd76c-kube-api-access-k8wg6\") pod \"collect-profiles-29322270-qb9dd\" (UID: \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.361617 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f684623e-fdca-4fa3-9b78-ef67b56cd76c-secret-volume\") pod \"collect-profiles-29322270-qb9dd\" (UID: \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.362449 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f684623e-fdca-4fa3-9b78-ef67b56cd76c-config-volume\") pod \"collect-profiles-29322270-qb9dd\" (UID: \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.369802 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f684623e-fdca-4fa3-9b78-ef67b56cd76c-secret-volume\") pod \"collect-profiles-29322270-qb9dd\" (UID: \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.385119 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8wg6\" (UniqueName: \"kubernetes.io/projected/f684623e-fdca-4fa3-9b78-ef67b56cd76c-kube-api-access-k8wg6\") pod \"collect-profiles-29322270-qb9dd\" (UID: \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" Oct 01 16:30:00 crc kubenswrapper[4726]: I1001 16:30:00.532673 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" Oct 01 16:30:01 crc kubenswrapper[4726]: I1001 16:30:01.058641 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd"] Oct 01 16:30:01 crc kubenswrapper[4726]: W1001 16:30:01.068780 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf684623e_fdca_4fa3_9b78_ef67b56cd76c.slice/crio-6ebe327ddcf54f5af237cc9066ad777afaf02d9445ab8e819864cd980e6c60ac WatchSource:0}: Error finding container 6ebe327ddcf54f5af237cc9066ad777afaf02d9445ab8e819864cd980e6c60ac: Status 404 returned error can't find the container with id 6ebe327ddcf54f5af237cc9066ad777afaf02d9445ab8e819864cd980e6c60ac Oct 01 16:30:01 crc kubenswrapper[4726]: I1001 16:30:01.338899 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" event={"ID":"f684623e-fdca-4fa3-9b78-ef67b56cd76c","Type":"ContainerStarted","Data":"1b415d07040942a5090e5c1113b4b1caaa32e7f777701397c50ba3af2613fd27"} Oct 01 16:30:01 crc kubenswrapper[4726]: I1001 16:30:01.339427 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" event={"ID":"f684623e-fdca-4fa3-9b78-ef67b56cd76c","Type":"ContainerStarted","Data":"6ebe327ddcf54f5af237cc9066ad777afaf02d9445ab8e819864cd980e6c60ac"} Oct 01 16:30:01 crc kubenswrapper[4726]: I1001 16:30:01.345481 4726 generic.go:334] "Generic (PLEG): container finished" podID="c24a10a9-6d60-43d5-9e3c-67cec3e0045c" containerID="ec83eb11966398c2c428a8cd4ac1354b44c182f3bcf9e1358f8d0a13781bd0db" exitCode=0 Oct 01 16:30:01 crc kubenswrapper[4726]: I1001 16:30:01.345555 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nj4wf" event={"ID":"c24a10a9-6d60-43d5-9e3c-67cec3e0045c","Type":"ContainerDied","Data":"ec83eb11966398c2c428a8cd4ac1354b44c182f3bcf9e1358f8d0a13781bd0db"} Oct 01 16:30:01 crc kubenswrapper[4726]: I1001 16:30:01.366906 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" podStartSLOduration=1.366888669 podStartE2EDuration="1.366888669s" podCreationTimestamp="2025-10-01 16:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:30:01.360176704 +0000 UTC m=+1854.261729291" watchObservedRunningTime="2025-10-01 16:30:01.366888669 +0000 UTC m=+1854.268441246" Oct 01 16:30:02 crc kubenswrapper[4726]: I1001 16:30:02.360329 4726 generic.go:334] "Generic (PLEG): container finished" podID="f684623e-fdca-4fa3-9b78-ef67b56cd76c" containerID="1b415d07040942a5090e5c1113b4b1caaa32e7f777701397c50ba3af2613fd27" exitCode=0 Oct 01 16:30:02 crc kubenswrapper[4726]: I1001 16:30:02.360402 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" event={"ID":"f684623e-fdca-4fa3-9b78-ef67b56cd76c","Type":"ContainerDied","Data":"1b415d07040942a5090e5c1113b4b1caaa32e7f777701397c50ba3af2613fd27"} Oct 01 16:30:03 crc kubenswrapper[4726]: I1001 16:30:03.687275 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" Oct 01 16:30:03 crc kubenswrapper[4726]: I1001 16:30:03.850986 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f684623e-fdca-4fa3-9b78-ef67b56cd76c-config-volume\") pod \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\" (UID: \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\") " Oct 01 16:30:03 crc kubenswrapper[4726]: I1001 16:30:03.851159 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8wg6\" (UniqueName: \"kubernetes.io/projected/f684623e-fdca-4fa3-9b78-ef67b56cd76c-kube-api-access-k8wg6\") pod \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\" (UID: \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\") " Oct 01 16:30:03 crc kubenswrapper[4726]: I1001 16:30:03.851316 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f684623e-fdca-4fa3-9b78-ef67b56cd76c-secret-volume\") pod \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\" (UID: \"f684623e-fdca-4fa3-9b78-ef67b56cd76c\") " Oct 01 16:30:03 crc kubenswrapper[4726]: I1001 16:30:03.852426 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f684623e-fdca-4fa3-9b78-ef67b56cd76c-config-volume" (OuterVolumeSpecName: "config-volume") pod "f684623e-fdca-4fa3-9b78-ef67b56cd76c" (UID: "f684623e-fdca-4fa3-9b78-ef67b56cd76c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:30:03 crc kubenswrapper[4726]: I1001 16:30:03.860313 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f684623e-fdca-4fa3-9b78-ef67b56cd76c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f684623e-fdca-4fa3-9b78-ef67b56cd76c" (UID: "f684623e-fdca-4fa3-9b78-ef67b56cd76c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:30:03 crc kubenswrapper[4726]: I1001 16:30:03.860396 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f684623e-fdca-4fa3-9b78-ef67b56cd76c-kube-api-access-k8wg6" (OuterVolumeSpecName: "kube-api-access-k8wg6") pod "f684623e-fdca-4fa3-9b78-ef67b56cd76c" (UID: "f684623e-fdca-4fa3-9b78-ef67b56cd76c"). InnerVolumeSpecName "kube-api-access-k8wg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:30:03 crc kubenswrapper[4726]: I1001 16:30:03.953171 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f684623e-fdca-4fa3-9b78-ef67b56cd76c-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:03 crc kubenswrapper[4726]: I1001 16:30:03.953210 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f684623e-fdca-4fa3-9b78-ef67b56cd76c-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:03 crc kubenswrapper[4726]: I1001 16:30:03.953223 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8wg6\" (UniqueName: \"kubernetes.io/projected/f684623e-fdca-4fa3-9b78-ef67b56cd76c-kube-api-access-k8wg6\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:04 crc kubenswrapper[4726]: I1001 16:30:04.378794 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" event={"ID":"f684623e-fdca-4fa3-9b78-ef67b56cd76c","Type":"ContainerDied","Data":"6ebe327ddcf54f5af237cc9066ad777afaf02d9445ab8e819864cd980e6c60ac"} Oct 01 16:30:04 crc kubenswrapper[4726]: I1001 16:30:04.379147 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ebe327ddcf54f5af237cc9066ad777afaf02d9445ab8e819864cd980e6c60ac" Oct 01 16:30:04 crc kubenswrapper[4726]: I1001 16:30:04.378859 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-qb9dd" Oct 01 16:30:11 crc kubenswrapper[4726]: I1001 16:30:11.809093 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:30:11 crc kubenswrapper[4726]: E1001 16:30:11.809870 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:30:12 crc kubenswrapper[4726]: I1001 16:30:12.449738 4726 generic.go:334] "Generic (PLEG): container finished" podID="98f0cdbd-7ea0-486b-92bc-14e78ab54ffd" containerID="0b1573d7011f4191feb6f9da54a5f346a031e71958691fa4e5304f1a6970c4eb" exitCode=2 Oct 01 16:30:12 crc kubenswrapper[4726]: I1001 16:30:12.449798 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" event={"ID":"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd","Type":"ContainerDied","Data":"0b1573d7011f4191feb6f9da54a5f346a031e71958691fa4e5304f1a6970c4eb"} Oct 01 16:30:13 crc kubenswrapper[4726]: I1001 16:30:13.466333 4726 generic.go:334] "Generic (PLEG): container finished" podID="c24a10a9-6d60-43d5-9e3c-67cec3e0045c" containerID="c64040d96f7bc3f6ff142139d92f588e37f00228051fc8b2482905be9d1970ec" exitCode=0 Oct 01 16:30:13 crc kubenswrapper[4726]: I1001 16:30:13.466474 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nj4wf" event={"ID":"c24a10a9-6d60-43d5-9e3c-67cec3e0045c","Type":"ContainerDied","Data":"c64040d96f7bc3f6ff142139d92f588e37f00228051fc8b2482905be9d1970ec"} Oct 01 16:30:13 crc kubenswrapper[4726]: I1001 16:30:13.902228 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" Oct 01 16:30:14 crc kubenswrapper[4726]: I1001 16:30:14.070042 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-ssh-key\") pod \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\" (UID: \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\") " Oct 01 16:30:14 crc kubenswrapper[4726]: I1001 16:30:14.070177 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qm9x7\" (UniqueName: \"kubernetes.io/projected/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-kube-api-access-qm9x7\") pod \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\" (UID: \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\") " Oct 01 16:30:14 crc kubenswrapper[4726]: I1001 16:30:14.070224 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-inventory\") pod \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\" (UID: \"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd\") " Oct 01 16:30:14 crc kubenswrapper[4726]: I1001 16:30:14.075931 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-kube-api-access-qm9x7" (OuterVolumeSpecName: "kube-api-access-qm9x7") pod "98f0cdbd-7ea0-486b-92bc-14e78ab54ffd" (UID: "98f0cdbd-7ea0-486b-92bc-14e78ab54ffd"). InnerVolumeSpecName "kube-api-access-qm9x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:30:14 crc kubenswrapper[4726]: I1001 16:30:14.097073 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "98f0cdbd-7ea0-486b-92bc-14e78ab54ffd" (UID: "98f0cdbd-7ea0-486b-92bc-14e78ab54ffd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:30:14 crc kubenswrapper[4726]: I1001 16:30:14.109750 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-inventory" (OuterVolumeSpecName: "inventory") pod "98f0cdbd-7ea0-486b-92bc-14e78ab54ffd" (UID: "98f0cdbd-7ea0-486b-92bc-14e78ab54ffd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:30:14 crc kubenswrapper[4726]: I1001 16:30:14.171626 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:14 crc kubenswrapper[4726]: I1001 16:30:14.171662 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:14 crc kubenswrapper[4726]: I1001 16:30:14.171675 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qm9x7\" (UniqueName: \"kubernetes.io/projected/98f0cdbd-7ea0-486b-92bc-14e78ab54ffd-kube-api-access-qm9x7\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:14 crc kubenswrapper[4726]: I1001 16:30:14.477690 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nj4wf" event={"ID":"c24a10a9-6d60-43d5-9e3c-67cec3e0045c","Type":"ContainerStarted","Data":"80ed90b7ab8fa0abf6334fc7bf8ee7415f140ed8c2dc15f78ee0c6445d9d91f6"} Oct 01 16:30:14 crc kubenswrapper[4726]: I1001 16:30:14.479659 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" event={"ID":"98f0cdbd-7ea0-486b-92bc-14e78ab54ffd","Type":"ContainerDied","Data":"c353e0b076595c221df764f247c170318f0f330af2299a2679214358d8a4f559"} Oct 01 16:30:14 crc kubenswrapper[4726]: I1001 16:30:14.479698 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c353e0b076595c221df764f247c170318f0f330af2299a2679214358d8a4f559" Oct 01 16:30:14 crc kubenswrapper[4726]: I1001 16:30:14.479748 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk" Oct 01 16:30:14 crc kubenswrapper[4726]: I1001 16:30:14.505279 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nj4wf" podStartSLOduration=2.910623669 podStartE2EDuration="15.505257158s" podCreationTimestamp="2025-10-01 16:29:59 +0000 UTC" firstStartedPulling="2025-10-01 16:30:01.34725634 +0000 UTC m=+1854.248808917" lastFinishedPulling="2025-10-01 16:30:13.941889829 +0000 UTC m=+1866.843442406" observedRunningTime="2025-10-01 16:30:14.495576408 +0000 UTC m=+1867.397129005" watchObservedRunningTime="2025-10-01 16:30:14.505257158 +0000 UTC m=+1867.406809745" Oct 01 16:30:14 crc kubenswrapper[4726]: E1001 16:30:14.558119 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98f0cdbd_7ea0_486b_92bc_14e78ab54ffd.slice\": RecentStats: unable to find data in memory cache]" Oct 01 16:30:19 crc kubenswrapper[4726]: I1001 16:30:19.870443 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:30:19 crc kubenswrapper[4726]: I1001 16:30:19.871291 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:30:19 crc kubenswrapper[4726]: I1001 16:30:19.919483 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:30:20 crc kubenswrapper[4726]: I1001 16:30:20.595200 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nj4wf" Oct 01 16:30:20 crc kubenswrapper[4726]: I1001 16:30:20.686321 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nj4wf"] Oct 01 16:30:20 crc kubenswrapper[4726]: I1001 16:30:20.748980 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2ql9b"] Oct 01 16:30:20 crc kubenswrapper[4726]: I1001 16:30:20.749388 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2ql9b" podUID="2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" containerName="registry-server" containerID="cri-o://a9c9ec66f518e18cec5e6c038d9b0a6ab8e4f31e36c0ad2368da7ce0f6f975a9" gracePeriod=2 Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.041942 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w"] Oct 01 16:30:21 crc kubenswrapper[4726]: E1001 16:30:21.042761 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f684623e-fdca-4fa3-9b78-ef67b56cd76c" containerName="collect-profiles" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.042780 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f684623e-fdca-4fa3-9b78-ef67b56cd76c" containerName="collect-profiles" Oct 01 16:30:21 crc kubenswrapper[4726]: E1001 16:30:21.042807 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f0cdbd-7ea0-486b-92bc-14e78ab54ffd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.042817 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f0cdbd-7ea0-486b-92bc-14e78ab54ffd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.043269 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="98f0cdbd-7ea0-486b-92bc-14e78ab54ffd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.043288 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f684623e-fdca-4fa3-9b78-ef67b56cd76c" containerName="collect-profiles" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.044184 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.052075 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.052080 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.052258 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.052281 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.069296 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w"] Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.217228 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c389e9-bf2c-4e85-a2ec-fc334262347f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4s76w\" (UID: \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.217527 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db4v8\" (UniqueName: \"kubernetes.io/projected/b8c389e9-bf2c-4e85-a2ec-fc334262347f-kube-api-access-db4v8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4s76w\" (UID: \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.217631 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8c389e9-bf2c-4e85-a2ec-fc334262347f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4s76w\" (UID: \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.319428 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c389e9-bf2c-4e85-a2ec-fc334262347f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4s76w\" (UID: \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.319514 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db4v8\" (UniqueName: \"kubernetes.io/projected/b8c389e9-bf2c-4e85-a2ec-fc334262347f-kube-api-access-db4v8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4s76w\" (UID: \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.319607 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8c389e9-bf2c-4e85-a2ec-fc334262347f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4s76w\" (UID: \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.331015 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c389e9-bf2c-4e85-a2ec-fc334262347f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4s76w\" (UID: \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.331042 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8c389e9-bf2c-4e85-a2ec-fc334262347f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4s76w\" (UID: \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.337617 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db4v8\" (UniqueName: \"kubernetes.io/projected/b8c389e9-bf2c-4e85-a2ec-fc334262347f-kube-api-access-db4v8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4s76w\" (UID: \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.395533 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.401113 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.522302 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-catalog-content\") pod \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\" (UID: \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\") " Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.522644 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc5lz\" (UniqueName: \"kubernetes.io/projected/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-kube-api-access-nc5lz\") pod \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\" (UID: \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\") " Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.522813 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-utilities\") pod \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\" (UID: \"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a\") " Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.525108 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-utilities" (OuterVolumeSpecName: "utilities") pod "2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" (UID: "2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.531917 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-kube-api-access-nc5lz" (OuterVolumeSpecName: "kube-api-access-nc5lz") pod "2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" (UID: "2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a"). InnerVolumeSpecName "kube-api-access-nc5lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.556137 4726 generic.go:334] "Generic (PLEG): container finished" podID="2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" containerID="a9c9ec66f518e18cec5e6c038d9b0a6ab8e4f31e36c0ad2368da7ce0f6f975a9" exitCode=0 Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.557120 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2ql9b" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.557143 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2ql9b" event={"ID":"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a","Type":"ContainerDied","Data":"a9c9ec66f518e18cec5e6c038d9b0a6ab8e4f31e36c0ad2368da7ce0f6f975a9"} Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.557207 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2ql9b" event={"ID":"2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a","Type":"ContainerDied","Data":"3d6a038380a7e43a6b91a88d289c58817ffbca92845046f646b9578ab1be4fd9"} Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.557226 4726 scope.go:117] "RemoveContainer" containerID="a9c9ec66f518e18cec5e6c038d9b0a6ab8e4f31e36c0ad2368da7ce0f6f975a9" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.583708 4726 scope.go:117] "RemoveContainer" containerID="fa1c4c8f3a84a3d7f2be578564919beda38dd35557284336a8c2d505a8cdc82b" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.601983 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" (UID: "2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.619191 4726 scope.go:117] "RemoveContainer" containerID="4f417a88ac4ad829e3698306fea3f478db1b51b95b6d4d23b5c2ea298f092f7b" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.626608 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.626633 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.626645 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc5lz\" (UniqueName: \"kubernetes.io/projected/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a-kube-api-access-nc5lz\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.680290 4726 scope.go:117] "RemoveContainer" containerID="a9c9ec66f518e18cec5e6c038d9b0a6ab8e4f31e36c0ad2368da7ce0f6f975a9" Oct 01 16:30:21 crc kubenswrapper[4726]: E1001 16:30:21.680809 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9c9ec66f518e18cec5e6c038d9b0a6ab8e4f31e36c0ad2368da7ce0f6f975a9\": container with ID starting with a9c9ec66f518e18cec5e6c038d9b0a6ab8e4f31e36c0ad2368da7ce0f6f975a9 not found: ID does not exist" containerID="a9c9ec66f518e18cec5e6c038d9b0a6ab8e4f31e36c0ad2368da7ce0f6f975a9" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.680845 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9c9ec66f518e18cec5e6c038d9b0a6ab8e4f31e36c0ad2368da7ce0f6f975a9"} err="failed to get container status \"a9c9ec66f518e18cec5e6c038d9b0a6ab8e4f31e36c0ad2368da7ce0f6f975a9\": rpc error: code = NotFound desc = could not find container \"a9c9ec66f518e18cec5e6c038d9b0a6ab8e4f31e36c0ad2368da7ce0f6f975a9\": container with ID starting with a9c9ec66f518e18cec5e6c038d9b0a6ab8e4f31e36c0ad2368da7ce0f6f975a9 not found: ID does not exist" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.680890 4726 scope.go:117] "RemoveContainer" containerID="fa1c4c8f3a84a3d7f2be578564919beda38dd35557284336a8c2d505a8cdc82b" Oct 01 16:30:21 crc kubenswrapper[4726]: E1001 16:30:21.683211 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa1c4c8f3a84a3d7f2be578564919beda38dd35557284336a8c2d505a8cdc82b\": container with ID starting with fa1c4c8f3a84a3d7f2be578564919beda38dd35557284336a8c2d505a8cdc82b not found: ID does not exist" containerID="fa1c4c8f3a84a3d7f2be578564919beda38dd35557284336a8c2d505a8cdc82b" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.683274 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa1c4c8f3a84a3d7f2be578564919beda38dd35557284336a8c2d505a8cdc82b"} err="failed to get container status \"fa1c4c8f3a84a3d7f2be578564919beda38dd35557284336a8c2d505a8cdc82b\": rpc error: code = NotFound desc = could not find container \"fa1c4c8f3a84a3d7f2be578564919beda38dd35557284336a8c2d505a8cdc82b\": container with ID starting with fa1c4c8f3a84a3d7f2be578564919beda38dd35557284336a8c2d505a8cdc82b not found: ID does not exist" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.683306 4726 scope.go:117] "RemoveContainer" containerID="4f417a88ac4ad829e3698306fea3f478db1b51b95b6d4d23b5c2ea298f092f7b" Oct 01 16:30:21 crc kubenswrapper[4726]: E1001 16:30:21.683632 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f417a88ac4ad829e3698306fea3f478db1b51b95b6d4d23b5c2ea298f092f7b\": container with ID starting with 4f417a88ac4ad829e3698306fea3f478db1b51b95b6d4d23b5c2ea298f092f7b not found: ID does not exist" containerID="4f417a88ac4ad829e3698306fea3f478db1b51b95b6d4d23b5c2ea298f092f7b" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.683685 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f417a88ac4ad829e3698306fea3f478db1b51b95b6d4d23b5c2ea298f092f7b"} err="failed to get container status \"4f417a88ac4ad829e3698306fea3f478db1b51b95b6d4d23b5c2ea298f092f7b\": rpc error: code = NotFound desc = could not find container \"4f417a88ac4ad829e3698306fea3f478db1b51b95b6d4d23b5c2ea298f092f7b\": container with ID starting with 4f417a88ac4ad829e3698306fea3f478db1b51b95b6d4d23b5c2ea298f092f7b not found: ID does not exist" Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.878278 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2ql9b"] Oct 01 16:30:21 crc kubenswrapper[4726]: I1001 16:30:21.886286 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2ql9b"] Oct 01 16:30:22 crc kubenswrapper[4726]: I1001 16:30:22.045746 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w"] Oct 01 16:30:22 crc kubenswrapper[4726]: I1001 16:30:22.565887 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" event={"ID":"b8c389e9-bf2c-4e85-a2ec-fc334262347f","Type":"ContainerStarted","Data":"3a19c533e4166de6016702ab42c88d76acdf82434b7fddb06cb3d00efa62e961"} Oct 01 16:30:22 crc kubenswrapper[4726]: I1001 16:30:22.808762 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:30:22 crc kubenswrapper[4726]: E1001 16:30:22.809359 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.305909 4726 scope.go:117] "RemoveContainer" containerID="04aa3a2c1c842a1fb997a6f412f4d5e2fa43fbf04dc54e8452e295344b5cf5ca" Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.392041 4726 scope.go:117] "RemoveContainer" containerID="ed81401bbb74ca249f20375c440e14f84f09d3a54734aef9dc6aed10e691944d" Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.587201 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" event={"ID":"b8c389e9-bf2c-4e85-a2ec-fc334262347f","Type":"ContainerStarted","Data":"11923a927cd267b9bde7ef93092ef6331c479aaae8fc9f8519e757a6f5ff33c6"} Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.612998 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" podStartSLOduration=2.055189167 podStartE2EDuration="2.612978705s" podCreationTimestamp="2025-10-01 16:30:21 +0000 UTC" firstStartedPulling="2025-10-01 16:30:22.051571958 +0000 UTC m=+1874.953124535" lastFinishedPulling="2025-10-01 16:30:22.609361496 +0000 UTC m=+1875.510914073" observedRunningTime="2025-10-01 16:30:23.606333713 +0000 UTC m=+1876.507886300" watchObservedRunningTime="2025-10-01 16:30:23.612978705 +0000 UTC m=+1876.514531282" Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.818046 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" path="/var/lib/kubelet/pods/2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a/volumes" Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.857745 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2"] Oct 01 16:30:23 crc kubenswrapper[4726]: E1001 16:30:23.858238 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" containerName="extract-content" Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.858262 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" containerName="extract-content" Oct 01 16:30:23 crc kubenswrapper[4726]: E1001 16:30:23.858287 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" containerName="extract-utilities" Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.858299 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" containerName="extract-utilities" Oct 01 16:30:23 crc kubenswrapper[4726]: E1001 16:30:23.858335 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" containerName="registry-server" Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.858344 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" containerName="registry-server" Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.858584 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dc72aec-f3f5-46b7-8fc7-aa160fa1f47a" containerName="registry-server" Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.860231 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.863537 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.877147 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2"] Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.970342 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2\" (UID: \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.972144 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2\" (UID: \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" Oct 01 16:30:23 crc kubenswrapper[4726]: I1001 16:30:23.973392 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwjxh\" (UniqueName: \"kubernetes.io/projected/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-kube-api-access-dwjxh\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2\" (UID: \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.007287 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s"] Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.009597 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.040504 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s"] Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.077121 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwjxh\" (UniqueName: \"kubernetes.io/projected/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-kube-api-access-dwjxh\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2\" (UID: \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.077197 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2\" (UID: \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.077284 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2\" (UID: \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.077778 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2\" (UID: \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.078052 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2\" (UID: \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.098380 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwjxh\" (UniqueName: \"kubernetes.io/projected/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-kube-api-access-dwjxh\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2\" (UID: \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.178911 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s\" (UID: \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.179088 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s\" (UID: \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.179225 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5vg8\" (UniqueName: \"kubernetes.io/projected/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-kube-api-access-w5vg8\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s\" (UID: \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.189395 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.281087 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5vg8\" (UniqueName: \"kubernetes.io/projected/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-kube-api-access-w5vg8\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s\" (UID: \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.281197 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s\" (UID: \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.281235 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s\" (UID: \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.281679 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s\" (UID: \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.282194 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s\" (UID: \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.302617 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5vg8\" (UniqueName: \"kubernetes.io/projected/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-kube-api-access-w5vg8\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s\" (UID: \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.329349 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.732960 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2"] Oct 01 16:30:24 crc kubenswrapper[4726]: I1001 16:30:24.886967 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s"] Oct 01 16:30:24 crc kubenswrapper[4726]: W1001 16:30:24.897380 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod213c0beb_44a8_4e18_afaa_4d2ba6fd6305.slice/crio-5e43cef455aff316f23cb05592eecfd7499e51d09ff1a532290b09968a97a807 WatchSource:0}: Error finding container 5e43cef455aff316f23cb05592eecfd7499e51d09ff1a532290b09968a97a807: Status 404 returned error can't find the container with id 5e43cef455aff316f23cb05592eecfd7499e51d09ff1a532290b09968a97a807 Oct 01 16:30:25 crc kubenswrapper[4726]: I1001 16:30:25.610470 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" event={"ID":"213c0beb-44a8-4e18-afaa-4d2ba6fd6305","Type":"ContainerStarted","Data":"5e43cef455aff316f23cb05592eecfd7499e51d09ff1a532290b09968a97a807"} Oct 01 16:30:25 crc kubenswrapper[4726]: I1001 16:30:25.613650 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" event={"ID":"7b3a32c8-9c47-4ccc-9885-d29009a22aa4","Type":"ContainerStarted","Data":"94107a816ce629f29e5128f0e781208db98f66249e469bcd1dd67625219e1603"} Oct 01 16:30:26 crc kubenswrapper[4726]: I1001 16:30:26.626502 4726 generic.go:334] "Generic (PLEG): container finished" podID="7b3a32c8-9c47-4ccc-9885-d29009a22aa4" containerID="85b9e9d430341c273e358baa4ff5a3291ad9e09ee2ecc2a29f6c1cbbf252d69b" exitCode=0 Oct 01 16:30:26 crc kubenswrapper[4726]: I1001 16:30:26.626588 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" event={"ID":"7b3a32c8-9c47-4ccc-9885-d29009a22aa4","Type":"ContainerDied","Data":"85b9e9d430341c273e358baa4ff5a3291ad9e09ee2ecc2a29f6c1cbbf252d69b"} Oct 01 16:30:26 crc kubenswrapper[4726]: I1001 16:30:26.628375 4726 generic.go:334] "Generic (PLEG): container finished" podID="213c0beb-44a8-4e18-afaa-4d2ba6fd6305" containerID="3cc0afb36109f8c0403ae53adabb1176909ebaf370a12fabad49d1a6ef0b7a10" exitCode=0 Oct 01 16:30:26 crc kubenswrapper[4726]: I1001 16:30:26.628420 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" event={"ID":"213c0beb-44a8-4e18-afaa-4d2ba6fd6305","Type":"ContainerDied","Data":"3cc0afb36109f8c0403ae53adabb1176909ebaf370a12fabad49d1a6ef0b7a10"} Oct 01 16:30:26 crc kubenswrapper[4726]: I1001 16:30:26.629237 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:30:28 crc kubenswrapper[4726]: I1001 16:30:28.647945 4726 generic.go:334] "Generic (PLEG): container finished" podID="213c0beb-44a8-4e18-afaa-4d2ba6fd6305" containerID="2826c82dc520882781c1ef573f15f668da97bf0647adb5ce30af70fd927dace1" exitCode=0 Oct 01 16:30:28 crc kubenswrapper[4726]: I1001 16:30:28.648013 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" event={"ID":"213c0beb-44a8-4e18-afaa-4d2ba6fd6305","Type":"ContainerDied","Data":"2826c82dc520882781c1ef573f15f668da97bf0647adb5ce30af70fd927dace1"} Oct 01 16:30:29 crc kubenswrapper[4726]: I1001 16:30:29.038179 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-vj74b"] Oct 01 16:30:29 crc kubenswrapper[4726]: I1001 16:30:29.052291 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-vj74b"] Oct 01 16:30:29 crc kubenswrapper[4726]: I1001 16:30:29.662388 4726 generic.go:334] "Generic (PLEG): container finished" podID="213c0beb-44a8-4e18-afaa-4d2ba6fd6305" containerID="74e47d11ba3503339856d600fd580cb4ad86dd687471c98c336da8571e79b254" exitCode=0 Oct 01 16:30:29 crc kubenswrapper[4726]: I1001 16:30:29.662466 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" event={"ID":"213c0beb-44a8-4e18-afaa-4d2ba6fd6305","Type":"ContainerDied","Data":"74e47d11ba3503339856d600fd580cb4ad86dd687471c98c336da8571e79b254"} Oct 01 16:30:29 crc kubenswrapper[4726]: I1001 16:30:29.665250 4726 generic.go:334] "Generic (PLEG): container finished" podID="7b3a32c8-9c47-4ccc-9885-d29009a22aa4" containerID="7188f5a6c6a7ee6b8a5f6f941c04b8387a4deca9a6788a530145dd07575b9f5f" exitCode=0 Oct 01 16:30:29 crc kubenswrapper[4726]: I1001 16:30:29.665322 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" event={"ID":"7b3a32c8-9c47-4ccc-9885-d29009a22aa4","Type":"ContainerDied","Data":"7188f5a6c6a7ee6b8a5f6f941c04b8387a4deca9a6788a530145dd07575b9f5f"} Oct 01 16:30:29 crc kubenswrapper[4726]: I1001 16:30:29.819841 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5033e1a8-ab77-4894-ac5b-95ca39e731fa" path="/var/lib/kubelet/pods/5033e1a8-ab77-4894-ac5b-95ca39e731fa/volumes" Oct 01 16:30:30 crc kubenswrapper[4726]: I1001 16:30:30.675500 4726 generic.go:334] "Generic (PLEG): container finished" podID="7b3a32c8-9c47-4ccc-9885-d29009a22aa4" containerID="8755fd565f7bedf1d4ee58471302c380b47942b031bc1b43b8e26d930c656ad7" exitCode=0 Oct 01 16:30:30 crc kubenswrapper[4726]: I1001 16:30:30.675565 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" event={"ID":"7b3a32c8-9c47-4ccc-9885-d29009a22aa4","Type":"ContainerDied","Data":"8755fd565f7bedf1d4ee58471302c380b47942b031bc1b43b8e26d930c656ad7"} Oct 01 16:30:31 crc kubenswrapper[4726]: I1001 16:30:31.027160 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" Oct 01 16:30:31 crc kubenswrapper[4726]: I1001 16:30:31.130118 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5vg8\" (UniqueName: \"kubernetes.io/projected/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-kube-api-access-w5vg8\") pod \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\" (UID: \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\") " Oct 01 16:30:31 crc kubenswrapper[4726]: I1001 16:30:31.130321 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-bundle\") pod \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\" (UID: \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\") " Oct 01 16:30:31 crc kubenswrapper[4726]: I1001 16:30:31.130361 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-util\") pod \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\" (UID: \"213c0beb-44a8-4e18-afaa-4d2ba6fd6305\") " Oct 01 16:30:31 crc kubenswrapper[4726]: I1001 16:30:31.131326 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-bundle" (OuterVolumeSpecName: "bundle") pod "213c0beb-44a8-4e18-afaa-4d2ba6fd6305" (UID: "213c0beb-44a8-4e18-afaa-4d2ba6fd6305"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:30:31 crc kubenswrapper[4726]: I1001 16:30:31.139127 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-kube-api-access-w5vg8" (OuterVolumeSpecName: "kube-api-access-w5vg8") pod "213c0beb-44a8-4e18-afaa-4d2ba6fd6305" (UID: "213c0beb-44a8-4e18-afaa-4d2ba6fd6305"). InnerVolumeSpecName "kube-api-access-w5vg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:30:31 crc kubenswrapper[4726]: I1001 16:30:31.233985 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5vg8\" (UniqueName: \"kubernetes.io/projected/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-kube-api-access-w5vg8\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:31 crc kubenswrapper[4726]: I1001 16:30:31.234035 4726 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:31 crc kubenswrapper[4726]: I1001 16:30:31.730242 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" Oct 01 16:30:31 crc kubenswrapper[4726]: I1001 16:30:31.730265 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s" event={"ID":"213c0beb-44a8-4e18-afaa-4d2ba6fd6305","Type":"ContainerDied","Data":"5e43cef455aff316f23cb05592eecfd7499e51d09ff1a532290b09968a97a807"} Oct 01 16:30:31 crc kubenswrapper[4726]: I1001 16:30:31.730779 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e43cef455aff316f23cb05592eecfd7499e51d09ff1a532290b09968a97a807" Oct 01 16:30:31 crc kubenswrapper[4726]: I1001 16:30:31.859984 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-util" (OuterVolumeSpecName: "util") pod "213c0beb-44a8-4e18-afaa-4d2ba6fd6305" (UID: "213c0beb-44a8-4e18-afaa-4d2ba6fd6305"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:30:31 crc kubenswrapper[4726]: I1001 16:30:31.946976 4726 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/213c0beb-44a8-4e18-afaa-4d2ba6fd6305-util\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:32 crc kubenswrapper[4726]: I1001 16:30:32.083025 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" Oct 01 16:30:32 crc kubenswrapper[4726]: I1001 16:30:32.251583 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-util\") pod \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\" (UID: \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\") " Oct 01 16:30:32 crc kubenswrapper[4726]: I1001 16:30:32.251740 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-bundle\") pod \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\" (UID: \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\") " Oct 01 16:30:32 crc kubenswrapper[4726]: I1001 16:30:32.251905 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwjxh\" (UniqueName: \"kubernetes.io/projected/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-kube-api-access-dwjxh\") pod \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\" (UID: \"7b3a32c8-9c47-4ccc-9885-d29009a22aa4\") " Oct 01 16:30:32 crc kubenswrapper[4726]: I1001 16:30:32.253831 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-bundle" (OuterVolumeSpecName: "bundle") pod "7b3a32c8-9c47-4ccc-9885-d29009a22aa4" (UID: "7b3a32c8-9c47-4ccc-9885-d29009a22aa4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:30:32 crc kubenswrapper[4726]: I1001 16:30:32.259373 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-kube-api-access-dwjxh" (OuterVolumeSpecName: "kube-api-access-dwjxh") pod "7b3a32c8-9c47-4ccc-9885-d29009a22aa4" (UID: "7b3a32c8-9c47-4ccc-9885-d29009a22aa4"). InnerVolumeSpecName "kube-api-access-dwjxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:30:32 crc kubenswrapper[4726]: I1001 16:30:32.264536 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-util" (OuterVolumeSpecName: "util") pod "7b3a32c8-9c47-4ccc-9885-d29009a22aa4" (UID: "7b3a32c8-9c47-4ccc-9885-d29009a22aa4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:30:32 crc kubenswrapper[4726]: I1001 16:30:32.354469 4726 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-util\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:32 crc kubenswrapper[4726]: I1001 16:30:32.354783 4726 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:32 crc kubenswrapper[4726]: I1001 16:30:32.354794 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwjxh\" (UniqueName: \"kubernetes.io/projected/7b3a32c8-9c47-4ccc-9885-d29009a22aa4-kube-api-access-dwjxh\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:32 crc kubenswrapper[4726]: I1001 16:30:32.746360 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" event={"ID":"7b3a32c8-9c47-4ccc-9885-d29009a22aa4","Type":"ContainerDied","Data":"94107a816ce629f29e5128f0e781208db98f66249e469bcd1dd67625219e1603"} Oct 01 16:30:32 crc kubenswrapper[4726]: I1001 16:30:32.746420 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94107a816ce629f29e5128f0e781208db98f66249e469bcd1dd67625219e1603" Oct 01 16:30:32 crc kubenswrapper[4726]: I1001 16:30:32.746461 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2" Oct 01 16:30:33 crc kubenswrapper[4726]: I1001 16:30:33.807931 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:30:34 crc kubenswrapper[4726]: I1001 16:30:34.767771 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"8ea6a11e2a728a74e08b777157550e1c40f0e83c9ab337602b7696ba954f563a"} Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.582112 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-mn7jt"] Oct 01 16:30:41 crc kubenswrapper[4726]: E1001 16:30:41.583015 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213c0beb-44a8-4e18-afaa-4d2ba6fd6305" containerName="pull" Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.583027 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="213c0beb-44a8-4e18-afaa-4d2ba6fd6305" containerName="pull" Oct 01 16:30:41 crc kubenswrapper[4726]: E1001 16:30:41.583044 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b3a32c8-9c47-4ccc-9885-d29009a22aa4" containerName="util" Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.583074 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b3a32c8-9c47-4ccc-9885-d29009a22aa4" containerName="util" Oct 01 16:30:41 crc kubenswrapper[4726]: E1001 16:30:41.583090 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213c0beb-44a8-4e18-afaa-4d2ba6fd6305" containerName="extract" Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.583097 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="213c0beb-44a8-4e18-afaa-4d2ba6fd6305" containerName="extract" Oct 01 16:30:41 crc kubenswrapper[4726]: E1001 16:30:41.583108 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213c0beb-44a8-4e18-afaa-4d2ba6fd6305" containerName="util" Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.583113 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="213c0beb-44a8-4e18-afaa-4d2ba6fd6305" containerName="util" Oct 01 16:30:41 crc kubenswrapper[4726]: E1001 16:30:41.583122 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b3a32c8-9c47-4ccc-9885-d29009a22aa4" containerName="extract" Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.583129 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b3a32c8-9c47-4ccc-9885-d29009a22aa4" containerName="extract" Oct 01 16:30:41 crc kubenswrapper[4726]: E1001 16:30:41.583143 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b3a32c8-9c47-4ccc-9885-d29009a22aa4" containerName="pull" Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.583149 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b3a32c8-9c47-4ccc-9885-d29009a22aa4" containerName="pull" Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.583315 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b3a32c8-9c47-4ccc-9885-d29009a22aa4" containerName="extract" Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.583333 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="213c0beb-44a8-4e18-afaa-4d2ba6fd6305" containerName="extract" Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.583911 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-mn7jt" Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.597629 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-mn7jt"] Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.745430 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgbsp\" (UniqueName: \"kubernetes.io/projected/d4e4c724-1025-48b6-a04c-77071be69269-kube-api-access-cgbsp\") pod \"nmstate-operator-858ddd8f98-mn7jt\" (UID: \"d4e4c724-1025-48b6-a04c-77071be69269\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-mn7jt" Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.847080 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgbsp\" (UniqueName: \"kubernetes.io/projected/d4e4c724-1025-48b6-a04c-77071be69269-kube-api-access-cgbsp\") pod \"nmstate-operator-858ddd8f98-mn7jt\" (UID: \"d4e4c724-1025-48b6-a04c-77071be69269\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-mn7jt" Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.872509 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgbsp\" (UniqueName: \"kubernetes.io/projected/d4e4c724-1025-48b6-a04c-77071be69269-kube-api-access-cgbsp\") pod \"nmstate-operator-858ddd8f98-mn7jt\" (UID: \"d4e4c724-1025-48b6-a04c-77071be69269\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-mn7jt" Oct 01 16:30:41 crc kubenswrapper[4726]: I1001 16:30:41.908084 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-mn7jt" Oct 01 16:30:42 crc kubenswrapper[4726]: I1001 16:30:42.373661 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-mn7jt"] Oct 01 16:30:42 crc kubenswrapper[4726]: I1001 16:30:42.838451 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-mn7jt" event={"ID":"d4e4c724-1025-48b6-a04c-77071be69269","Type":"ContainerStarted","Data":"3a43c71b2020ff046f56944a35aaf45be54af162a1663cdfeeff1b1404da05c2"} Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.369652 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz"] Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.370452 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" podUID="f179fe18-0a34-43df-bb39-dab69a9ee29b" containerName="nmstate-metrics" containerID="cri-o://15fff174bbe0d667779152131358e8d5810709cbfaa917040c43a8ad63a90369" gracePeriod=30 Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.370487 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" podUID="f179fe18-0a34-43df-bb39-dab69a9ee29b" containerName="kube-rbac-proxy" containerID="cri-o://41349d4807fad2f1eb449c5da33539c938175fdc87d92111b1019282e17302a8" gracePeriod=30 Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.391371 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-zftsg"] Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.391589 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" podUID="4c168234-66ba-492c-986c-835301bd0275" containerName="nmstate-webhook" containerID="cri-o://df28c3773d7ed079f09aceea751eca32bac74851679679cfa0d36e0a15abef2e" gracePeriod=30 Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.412387 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-handler-bg7hk"] Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.412600 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-nmstate/nmstate-handler-bg7hk" podUID="98229bdb-bd95-46a7-9b03-04698c7f2475" containerName="nmstate-handler" containerID="cri-o://09ccffad3d0057482a969cd85fd8ae0902881bb7552b904cc6d0aa11df4a7da8" gracePeriod=30 Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.523450 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm"] Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.530901 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.553242 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm"] Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.665670 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5fb896f2-8fbc-40d7-adf4-2a93089171ce-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-vqzfm\" (UID: \"5fb896f2-8fbc-40d7-adf4-2a93089171ce\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.665797 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fb896f2-8fbc-40d7-adf4-2a93089171ce-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-vqzfm\" (UID: \"5fb896f2-8fbc-40d7-adf4-2a93089171ce\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.665821 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkhzv\" (UniqueName: \"kubernetes.io/projected/5fb896f2-8fbc-40d7-adf4-2a93089171ce-kube-api-access-hkhzv\") pod \"nmstate-console-plugin-6b874cbd85-vqzfm\" (UID: \"5fb896f2-8fbc-40d7-adf4-2a93089171ce\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.747186 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.767726 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5fb896f2-8fbc-40d7-adf4-2a93089171ce-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-vqzfm\" (UID: \"5fb896f2-8fbc-40d7-adf4-2a93089171ce\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.768874 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fb896f2-8fbc-40d7-adf4-2a93089171ce-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-vqzfm\" (UID: \"5fb896f2-8fbc-40d7-adf4-2a93089171ce\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.768981 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkhzv\" (UniqueName: \"kubernetes.io/projected/5fb896f2-8fbc-40d7-adf4-2a93089171ce-kube-api-access-hkhzv\") pod \"nmstate-console-plugin-6b874cbd85-vqzfm\" (UID: \"5fb896f2-8fbc-40d7-adf4-2a93089171ce\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.769775 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fb896f2-8fbc-40d7-adf4-2a93089171ce-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-vqzfm\" (UID: \"5fb896f2-8fbc-40d7-adf4-2a93089171ce\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.795405 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5fb896f2-8fbc-40d7-adf4-2a93089171ce-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-vqzfm\" (UID: \"5fb896f2-8fbc-40d7-adf4-2a93089171ce\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.816509 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-kzwz4"] Oct 01 16:30:46 crc kubenswrapper[4726]: E1001 16:30:46.817004 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98229bdb-bd95-46a7-9b03-04698c7f2475" containerName="nmstate-handler" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.817020 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="98229bdb-bd95-46a7-9b03-04698c7f2475" containerName="nmstate-handler" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.817288 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="98229bdb-bd95-46a7-9b03-04698c7f2475" containerName="nmstate-handler" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.817410 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkhzv\" (UniqueName: \"kubernetes.io/projected/5fb896f2-8fbc-40d7-adf4-2a93089171ce-kube-api-access-hkhzv\") pod \"nmstate-console-plugin-6b874cbd85-vqzfm\" (UID: \"5fb896f2-8fbc-40d7-adf4-2a93089171ce\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.818002 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.870888 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-275r5\" (UniqueName: \"kubernetes.io/projected/98229bdb-bd95-46a7-9b03-04698c7f2475-kube-api-access-275r5\") pod \"98229bdb-bd95-46a7-9b03-04698c7f2475\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.870966 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-nmstate-lock\") pod \"98229bdb-bd95-46a7-9b03-04698c7f2475\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.870986 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-dbus-socket\") pod \"98229bdb-bd95-46a7-9b03-04698c7f2475\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.871148 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-ovs-socket\") pod \"98229bdb-bd95-46a7-9b03-04698c7f2475\" (UID: \"98229bdb-bd95-46a7-9b03-04698c7f2475\") " Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.871704 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-dbus-socket" (OuterVolumeSpecName: "dbus-socket") pod "98229bdb-bd95-46a7-9b03-04698c7f2475" (UID: "98229bdb-bd95-46a7-9b03-04698c7f2475"). InnerVolumeSpecName "dbus-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.871741 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-nmstate-lock" (OuterVolumeSpecName: "nmstate-lock") pod "98229bdb-bd95-46a7-9b03-04698c7f2475" (UID: "98229bdb-bd95-46a7-9b03-04698c7f2475"). InnerVolumeSpecName "nmstate-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.871760 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-ovs-socket" (OuterVolumeSpecName: "ovs-socket") pod "98229bdb-bd95-46a7-9b03-04698c7f2475" (UID: "98229bdb-bd95-46a7-9b03-04698c7f2475"). InnerVolumeSpecName "ovs-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.886200 4726 generic.go:334] "Generic (PLEG): container finished" podID="f179fe18-0a34-43df-bb39-dab69a9ee29b" containerID="41349d4807fad2f1eb449c5da33539c938175fdc87d92111b1019282e17302a8" exitCode=0 Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.886232 4726 generic.go:334] "Generic (PLEG): container finished" podID="f179fe18-0a34-43df-bb39-dab69a9ee29b" containerID="15fff174bbe0d667779152131358e8d5810709cbfaa917040c43a8ad63a90369" exitCode=0 Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.886338 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" event={"ID":"f179fe18-0a34-43df-bb39-dab69a9ee29b","Type":"ContainerDied","Data":"41349d4807fad2f1eb449c5da33539c938175fdc87d92111b1019282e17302a8"} Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.886376 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" event={"ID":"f179fe18-0a34-43df-bb39-dab69a9ee29b","Type":"ContainerDied","Data":"15fff174bbe0d667779152131358e8d5810709cbfaa917040c43a8ad63a90369"} Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.890967 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98229bdb-bd95-46a7-9b03-04698c7f2475-kube-api-access-275r5" (OuterVolumeSpecName: "kube-api-access-275r5") pod "98229bdb-bd95-46a7-9b03-04698c7f2475" (UID: "98229bdb-bd95-46a7-9b03-04698c7f2475"). InnerVolumeSpecName "kube-api-access-275r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.891202 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-mn7jt" event={"ID":"d4e4c724-1025-48b6-a04c-77071be69269","Type":"ContainerStarted","Data":"bd5bb8d1a0b86397650d544a752b9177bd704c274c22f3bde0bb6ded85d51c9d"} Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.894290 4726 generic.go:334] "Generic (PLEG): container finished" podID="98229bdb-bd95-46a7-9b03-04698c7f2475" containerID="09ccffad3d0057482a969cd85fd8ae0902881bb7552b904cc6d0aa11df4a7da8" exitCode=0 Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.894354 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-bg7hk" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.894484 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-bg7hk" event={"ID":"98229bdb-bd95-46a7-9b03-04698c7f2475","Type":"ContainerDied","Data":"09ccffad3d0057482a969cd85fd8ae0902881bb7552b904cc6d0aa11df4a7da8"} Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.894516 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-bg7hk" event={"ID":"98229bdb-bd95-46a7-9b03-04698c7f2475","Type":"ContainerDied","Data":"7dab191f6b1d9e19fd289b3c01888b708ab9e6422c6dc51c08dad7acdac612b4"} Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.894537 4726 scope.go:117] "RemoveContainer" containerID="09ccffad3d0057482a969cd85fd8ae0902881bb7552b904cc6d0aa11df4a7da8" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.898860 4726 generic.go:334] "Generic (PLEG): container finished" podID="4c168234-66ba-492c-986c-835301bd0275" containerID="df28c3773d7ed079f09aceea751eca32bac74851679679cfa0d36e0a15abef2e" exitCode=0 Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.898903 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" event={"ID":"4c168234-66ba-492c-986c-835301bd0275","Type":"ContainerDied","Data":"df28c3773d7ed079f09aceea751eca32bac74851679679cfa0d36e0a15abef2e"} Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.954737 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-mn7jt" podStartSLOduration=2.6882838319999998 podStartE2EDuration="5.95471163s" podCreationTimestamp="2025-10-01 16:30:41 +0000 UTC" firstStartedPulling="2025-10-01 16:30:42.381326826 +0000 UTC m=+1895.282879403" lastFinishedPulling="2025-10-01 16:30:45.647754624 +0000 UTC m=+1898.549307201" observedRunningTime="2025-10-01 16:30:46.928129319 +0000 UTC m=+1899.829681916" watchObservedRunningTime="2025-10-01 16:30:46.95471163 +0000 UTC m=+1899.856264217" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.963961 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8"] Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.964202 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8" podUID="5de9f82a-fe6f-47a8-be2a-0a513407b081" containerName="nmstate-operator" containerID="cri-o://f3626a99bbcbca80cdd89a6cbd66f2c1b6cadb8839c400ebf05eef48c1350422" gracePeriod=30 Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.975074 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6jb8\" (UniqueName: \"kubernetes.io/projected/fc59e6f9-76eb-49fb-9028-9e64820de406-kube-api-access-s6jb8\") pod \"nmstate-handler-kzwz4\" (UID: \"fc59e6f9-76eb-49fb-9028-9e64820de406\") " pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.975125 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/fc59e6f9-76eb-49fb-9028-9e64820de406-dbus-socket\") pod \"nmstate-handler-kzwz4\" (UID: \"fc59e6f9-76eb-49fb-9028-9e64820de406\") " pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.975248 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/fc59e6f9-76eb-49fb-9028-9e64820de406-ovs-socket\") pod \"nmstate-handler-kzwz4\" (UID: \"fc59e6f9-76eb-49fb-9028-9e64820de406\") " pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.975295 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/fc59e6f9-76eb-49fb-9028-9e64820de406-nmstate-lock\") pod \"nmstate-handler-kzwz4\" (UID: \"fc59e6f9-76eb-49fb-9028-9e64820de406\") " pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.975358 4726 reconciler_common.go:293] "Volume detached for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-ovs-socket\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.975374 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-275r5\" (UniqueName: \"kubernetes.io/projected/98229bdb-bd95-46a7-9b03-04698c7f2475-kube-api-access-275r5\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.975383 4726 reconciler_common.go:293] "Volume detached for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-nmstate-lock\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:46 crc kubenswrapper[4726]: I1001 16:30:46.975391 4726 reconciler_common.go:293] "Volume detached for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/98229bdb-bd95-46a7-9b03-04698c7f2475-dbus-socket\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.014116 4726 scope.go:117] "RemoveContainer" containerID="09ccffad3d0057482a969cd85fd8ae0902881bb7552b904cc6d0aa11df4a7da8" Oct 01 16:30:47 crc kubenswrapper[4726]: E1001 16:30:47.014520 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09ccffad3d0057482a969cd85fd8ae0902881bb7552b904cc6d0aa11df4a7da8\": container with ID starting with 09ccffad3d0057482a969cd85fd8ae0902881bb7552b904cc6d0aa11df4a7da8 not found: ID does not exist" containerID="09ccffad3d0057482a969cd85fd8ae0902881bb7552b904cc6d0aa11df4a7da8" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.014549 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09ccffad3d0057482a969cd85fd8ae0902881bb7552b904cc6d0aa11df4a7da8"} err="failed to get container status \"09ccffad3d0057482a969cd85fd8ae0902881bb7552b904cc6d0aa11df4a7da8\": rpc error: code = NotFound desc = could not find container \"09ccffad3d0057482a969cd85fd8ae0902881bb7552b904cc6d0aa11df4a7da8\": container with ID starting with 09ccffad3d0057482a969cd85fd8ae0902881bb7552b904cc6d0aa11df4a7da8 not found: ID does not exist" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.037349 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.076779 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/fc59e6f9-76eb-49fb-9028-9e64820de406-nmstate-lock\") pod \"nmstate-handler-kzwz4\" (UID: \"fc59e6f9-76eb-49fb-9028-9e64820de406\") " pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.076869 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6jb8\" (UniqueName: \"kubernetes.io/projected/fc59e6f9-76eb-49fb-9028-9e64820de406-kube-api-access-s6jb8\") pod \"nmstate-handler-kzwz4\" (UID: \"fc59e6f9-76eb-49fb-9028-9e64820de406\") " pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.076886 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/fc59e6f9-76eb-49fb-9028-9e64820de406-nmstate-lock\") pod \"nmstate-handler-kzwz4\" (UID: \"fc59e6f9-76eb-49fb-9028-9e64820de406\") " pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.076912 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/fc59e6f9-76eb-49fb-9028-9e64820de406-dbus-socket\") pod \"nmstate-handler-kzwz4\" (UID: \"fc59e6f9-76eb-49fb-9028-9e64820de406\") " pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.077478 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/fc59e6f9-76eb-49fb-9028-9e64820de406-ovs-socket\") pod \"nmstate-handler-kzwz4\" (UID: \"fc59e6f9-76eb-49fb-9028-9e64820de406\") " pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.078099 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/fc59e6f9-76eb-49fb-9028-9e64820de406-dbus-socket\") pod \"nmstate-handler-kzwz4\" (UID: \"fc59e6f9-76eb-49fb-9028-9e64820de406\") " pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.080093 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/fc59e6f9-76eb-49fb-9028-9e64820de406-ovs-socket\") pod \"nmstate-handler-kzwz4\" (UID: \"fc59e6f9-76eb-49fb-9028-9e64820de406\") " pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.093251 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6jb8\" (UniqueName: \"kubernetes.io/projected/fc59e6f9-76eb-49fb-9028-9e64820de406-kube-api-access-s6jb8\") pod \"nmstate-handler-kzwz4\" (UID: \"fc59e6f9-76eb-49fb-9028-9e64820de406\") " pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.110473 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.111520 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.127313 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-handler-bg7hk"] Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.127983 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.149776 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-nmstate/nmstate-handler-bg7hk"] Oct 01 16:30:47 crc kubenswrapper[4726]: W1001 16:30:47.172671 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc59e6f9_76eb_49fb_9028_9e64820de406.slice/crio-3aa62175038073f0a858a20a51439daa0548a515fbd39d104301d52f055cbf7b WatchSource:0}: Error finding container 3aa62175038073f0a858a20a51439daa0548a515fbd39d104301d52f055cbf7b: Status 404 returned error can't find the container with id 3aa62175038073f0a858a20a51439daa0548a515fbd39d104301d52f055cbf7b Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.288976 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bkh9\" (UniqueName: \"kubernetes.io/projected/4c168234-66ba-492c-986c-835301bd0275-kube-api-access-4bkh9\") pod \"4c168234-66ba-492c-986c-835301bd0275\" (UID: \"4c168234-66ba-492c-986c-835301bd0275\") " Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.289034 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4c168234-66ba-492c-986c-835301bd0275-tls-key-pair\") pod \"4c168234-66ba-492c-986c-835301bd0275\" (UID: \"4c168234-66ba-492c-986c-835301bd0275\") " Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.289316 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzhfn\" (UniqueName: \"kubernetes.io/projected/f179fe18-0a34-43df-bb39-dab69a9ee29b-kube-api-access-fzhfn\") pod \"f179fe18-0a34-43df-bb39-dab69a9ee29b\" (UID: \"f179fe18-0a34-43df-bb39-dab69a9ee29b\") " Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.294961 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2"] Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.295229 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c168234-66ba-492c-986c-835301bd0275-tls-key-pair" (OuterVolumeSpecName: "tls-key-pair") pod "4c168234-66ba-492c-986c-835301bd0275" (UID: "4c168234-66ba-492c-986c-835301bd0275"). InnerVolumeSpecName "tls-key-pair". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.295457 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f179fe18-0a34-43df-bb39-dab69a9ee29b-kube-api-access-fzhfn" (OuterVolumeSpecName: "kube-api-access-fzhfn") pod "f179fe18-0a34-43df-bb39-dab69a9ee29b" (UID: "f179fe18-0a34-43df-bb39-dab69a9ee29b"). InnerVolumeSpecName "kube-api-access-fzhfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:30:47 crc kubenswrapper[4726]: E1001 16:30:47.295652 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f179fe18-0a34-43df-bb39-dab69a9ee29b" containerName="nmstate-metrics" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.295674 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f179fe18-0a34-43df-bb39-dab69a9ee29b" containerName="nmstate-metrics" Oct 01 16:30:47 crc kubenswrapper[4726]: E1001 16:30:47.295697 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f179fe18-0a34-43df-bb39-dab69a9ee29b" containerName="kube-rbac-proxy" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.295706 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f179fe18-0a34-43df-bb39-dab69a9ee29b" containerName="kube-rbac-proxy" Oct 01 16:30:47 crc kubenswrapper[4726]: E1001 16:30:47.295767 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c168234-66ba-492c-986c-835301bd0275" containerName="nmstate-webhook" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.295777 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c168234-66ba-492c-986c-835301bd0275" containerName="nmstate-webhook" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.296102 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f179fe18-0a34-43df-bb39-dab69a9ee29b" containerName="nmstate-metrics" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.296142 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c168234-66ba-492c-986c-835301bd0275" containerName="nmstate-webhook" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.296158 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f179fe18-0a34-43df-bb39-dab69a9ee29b" containerName="kube-rbac-proxy" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.297232 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.302602 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2"] Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.335269 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c168234-66ba-492c-986c-835301bd0275-kube-api-access-4bkh9" (OuterVolumeSpecName: "kube-api-access-4bkh9") pod "4c168234-66ba-492c-986c-835301bd0275" (UID: "4c168234-66ba-492c-986c-835301bd0275"). InnerVolumeSpecName "kube-api-access-4bkh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.394034 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qplqh\" (UniqueName: \"kubernetes.io/projected/0f80c25b-98a7-46d0-bf07-26c19f885781-kube-api-access-qplqh\") pod \"nmstate-webhook-6cdbc54649-qrkz2\" (UID: \"0f80c25b-98a7-46d0-bf07-26c19f885781\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.394436 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/0f80c25b-98a7-46d0-bf07-26c19f885781-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-qrkz2\" (UID: \"0f80c25b-98a7-46d0-bf07-26c19f885781\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.394498 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bkh9\" (UniqueName: \"kubernetes.io/projected/4c168234-66ba-492c-986c-835301bd0275-kube-api-access-4bkh9\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.394512 4726 reconciler_common.go:293] "Volume detached for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4c168234-66ba-492c-986c-835301bd0275-tls-key-pair\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.394523 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzhfn\" (UniqueName: \"kubernetes.io/projected/f179fe18-0a34-43df-bb39-dab69a9ee29b-kube-api-access-fzhfn\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.399611 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-xxpvz"] Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.401307 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xxpvz" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.413529 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-xxpvz"] Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.497571 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b792t\" (UniqueName: \"kubernetes.io/projected/d24b198c-db7b-4057-bed1-6f32d0bac5d3-kube-api-access-b792t\") pod \"nmstate-metrics-fdff9cb8d-xxpvz\" (UID: \"d24b198c-db7b-4057-bed1-6f32d0bac5d3\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xxpvz" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.497707 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qplqh\" (UniqueName: \"kubernetes.io/projected/0f80c25b-98a7-46d0-bf07-26c19f885781-kube-api-access-qplqh\") pod \"nmstate-webhook-6cdbc54649-qrkz2\" (UID: \"0f80c25b-98a7-46d0-bf07-26c19f885781\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.498633 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/0f80c25b-98a7-46d0-bf07-26c19f885781-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-qrkz2\" (UID: \"0f80c25b-98a7-46d0-bf07-26c19f885781\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.506708 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/0f80c25b-98a7-46d0-bf07-26c19f885781-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-qrkz2\" (UID: \"0f80c25b-98a7-46d0-bf07-26c19f885781\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.507174 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.522883 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qplqh\" (UniqueName: \"kubernetes.io/projected/0f80c25b-98a7-46d0-bf07-26c19f885781-kube-api-access-qplqh\") pod \"nmstate-webhook-6cdbc54649-qrkz2\" (UID: \"0f80c25b-98a7-46d0-bf07-26c19f885781\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.599688 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtq2t\" (UniqueName: \"kubernetes.io/projected/5de9f82a-fe6f-47a8-be2a-0a513407b081-kube-api-access-rtq2t\") pod \"5de9f82a-fe6f-47a8-be2a-0a513407b081\" (UID: \"5de9f82a-fe6f-47a8-be2a-0a513407b081\") " Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.600208 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b792t\" (UniqueName: \"kubernetes.io/projected/d24b198c-db7b-4057-bed1-6f32d0bac5d3-kube-api-access-b792t\") pod \"nmstate-metrics-fdff9cb8d-xxpvz\" (UID: \"d24b198c-db7b-4057-bed1-6f32d0bac5d3\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xxpvz" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.602742 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5de9f82a-fe6f-47a8-be2a-0a513407b081-kube-api-access-rtq2t" (OuterVolumeSpecName: "kube-api-access-rtq2t") pod "5de9f82a-fe6f-47a8-be2a-0a513407b081" (UID: "5de9f82a-fe6f-47a8-be2a-0a513407b081"). InnerVolumeSpecName "kube-api-access-rtq2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.615601 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b792t\" (UniqueName: \"kubernetes.io/projected/d24b198c-db7b-4057-bed1-6f32d0bac5d3-kube-api-access-b792t\") pod \"nmstate-metrics-fdff9cb8d-xxpvz\" (UID: \"d24b198c-db7b-4057-bed1-6f32d0bac5d3\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xxpvz" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.664410 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm"] Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.665383 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2" Oct 01 16:30:47 crc kubenswrapper[4726]: W1001 16:30:47.666271 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fb896f2_8fbc_40d7_adf4_2a93089171ce.slice/crio-9f232efae4baff22e5c26cf7b00a27b8c3bf6d3b6b04593d853cbf23b862a5ab WatchSource:0}: Error finding container 9f232efae4baff22e5c26cf7b00a27b8c3bf6d3b6b04593d853cbf23b862a5ab: Status 404 returned error can't find the container with id 9f232efae4baff22e5c26cf7b00a27b8c3bf6d3b6b04593d853cbf23b862a5ab Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.701645 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtq2t\" (UniqueName: \"kubernetes.io/projected/5de9f82a-fe6f-47a8-be2a-0a513407b081-kube-api-access-rtq2t\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.726157 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xxpvz" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.852314 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98229bdb-bd95-46a7-9b03-04698c7f2475" path="/var/lib/kubelet/pods/98229bdb-bd95-46a7-9b03-04698c7f2475/volumes" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.926664 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm" event={"ID":"5fb896f2-8fbc-40d7-adf4-2a93089171ce","Type":"ContainerStarted","Data":"9f232efae4baff22e5c26cf7b00a27b8c3bf6d3b6b04593d853cbf23b862a5ab"} Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.939761 4726 generic.go:334] "Generic (PLEG): container finished" podID="5de9f82a-fe6f-47a8-be2a-0a513407b081" containerID="f3626a99bbcbca80cdd89a6cbd66f2c1b6cadb8839c400ebf05eef48c1350422" exitCode=0 Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.939811 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8" event={"ID":"5de9f82a-fe6f-47a8-be2a-0a513407b081","Type":"ContainerDied","Data":"f3626a99bbcbca80cdd89a6cbd66f2c1b6cadb8839c400ebf05eef48c1350422"} Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.939831 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8" event={"ID":"5de9f82a-fe6f-47a8-be2a-0a513407b081","Type":"ContainerDied","Data":"2f631e978f289cf0f0b69526f985aef51d8dbd2f909c18a06299a035e892f63f"} Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.939847 4726 scope.go:117] "RemoveContainer" containerID="f3626a99bbcbca80cdd89a6cbd66f2c1b6cadb8839c400ebf05eef48c1350422" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.939961 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.948989 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" event={"ID":"4c168234-66ba-492c-986c-835301bd0275","Type":"ContainerDied","Data":"95d3ce4a10223541ca661d7d9592c3f3f9f461ca4bf4d06769171ba50bd08718"} Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.949111 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-zftsg" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.968453 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" event={"ID":"f179fe18-0a34-43df-bb39-dab69a9ee29b","Type":"ContainerDied","Data":"e0898ee0e6c932ae4b7ae6523d8d2c16d1455fbe37b5faa1de588406ec3c4159"} Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.968562 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz" Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.979109 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8"] Oct 01 16:30:47 crc kubenswrapper[4726]: I1001 16:30:47.981675 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-kzwz4" event={"ID":"fc59e6f9-76eb-49fb-9028-9e64820de406","Type":"ContainerStarted","Data":"3aa62175038073f0a858a20a51439daa0548a515fbd39d104301d52f055cbf7b"} Oct 01 16:30:48 crc kubenswrapper[4726]: I1001 16:30:48.006920 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-fk9n8"] Oct 01 16:30:48 crc kubenswrapper[4726]: I1001 16:30:48.008440 4726 scope.go:117] "RemoveContainer" containerID="f3626a99bbcbca80cdd89a6cbd66f2c1b6cadb8839c400ebf05eef48c1350422" Oct 01 16:30:48 crc kubenswrapper[4726]: E1001 16:30:48.009638 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3626a99bbcbca80cdd89a6cbd66f2c1b6cadb8839c400ebf05eef48c1350422\": container with ID starting with f3626a99bbcbca80cdd89a6cbd66f2c1b6cadb8839c400ebf05eef48c1350422 not found: ID does not exist" containerID="f3626a99bbcbca80cdd89a6cbd66f2c1b6cadb8839c400ebf05eef48c1350422" Oct 01 16:30:48 crc kubenswrapper[4726]: I1001 16:30:48.009683 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3626a99bbcbca80cdd89a6cbd66f2c1b6cadb8839c400ebf05eef48c1350422"} err="failed to get container status \"f3626a99bbcbca80cdd89a6cbd66f2c1b6cadb8839c400ebf05eef48c1350422\": rpc error: code = NotFound desc = could not find container \"f3626a99bbcbca80cdd89a6cbd66f2c1b6cadb8839c400ebf05eef48c1350422\": container with ID starting with f3626a99bbcbca80cdd89a6cbd66f2c1b6cadb8839c400ebf05eef48c1350422 not found: ID does not exist" Oct 01 16:30:48 crc kubenswrapper[4726]: I1001 16:30:48.009710 4726 scope.go:117] "RemoveContainer" containerID="df28c3773d7ed079f09aceea751eca32bac74851679679cfa0d36e0a15abef2e" Oct 01 16:30:48 crc kubenswrapper[4726]: I1001 16:30:48.018014 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-zftsg"] Oct 01 16:30:48 crc kubenswrapper[4726]: I1001 16:30:48.040900 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-zftsg"] Oct 01 16:30:48 crc kubenswrapper[4726]: I1001 16:30:48.042891 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz"] Oct 01 16:30:48 crc kubenswrapper[4726]: I1001 16:30:48.051165 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-mdcjz"] Oct 01 16:30:48 crc kubenswrapper[4726]: I1001 16:30:48.065919 4726 scope.go:117] "RemoveContainer" containerID="41349d4807fad2f1eb449c5da33539c938175fdc87d92111b1019282e17302a8" Oct 01 16:30:48 crc kubenswrapper[4726]: I1001 16:30:48.085217 4726 scope.go:117] "RemoveContainer" containerID="15fff174bbe0d667779152131358e8d5810709cbfaa917040c43a8ad63a90369" Oct 01 16:30:48 crc kubenswrapper[4726]: I1001 16:30:48.182953 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2"] Oct 01 16:30:48 crc kubenswrapper[4726]: W1001 16:30:48.193236 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f80c25b_98a7_46d0_bf07_26c19f885781.slice/crio-21a54192dcde69365511cc0c6cbc2b997e44d86061fb138b89cb8c240a849bdb WatchSource:0}: Error finding container 21a54192dcde69365511cc0c6cbc2b997e44d86061fb138b89cb8c240a849bdb: Status 404 returned error can't find the container with id 21a54192dcde69365511cc0c6cbc2b997e44d86061fb138b89cb8c240a849bdb Oct 01 16:30:48 crc kubenswrapper[4726]: I1001 16:30:48.313618 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-xxpvz"] Oct 01 16:30:48 crc kubenswrapper[4726]: W1001 16:30:48.315502 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd24b198c_db7b_4057_bed1_6f32d0bac5d3.slice/crio-a5d276a9cd5fc31cb83880d0f29eca6afe898e83caa953813d61c62325feae66 WatchSource:0}: Error finding container a5d276a9cd5fc31cb83880d0f29eca6afe898e83caa953813d61c62325feae66: Status 404 returned error can't find the container with id a5d276a9cd5fc31cb83880d0f29eca6afe898e83caa953813d61c62325feae66 Oct 01 16:30:48 crc kubenswrapper[4726]: I1001 16:30:48.991469 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xxpvz" event={"ID":"d24b198c-db7b-4057-bed1-6f32d0bac5d3","Type":"ContainerStarted","Data":"a5d276a9cd5fc31cb83880d0f29eca6afe898e83caa953813d61c62325feae66"} Oct 01 16:30:48 crc kubenswrapper[4726]: I1001 16:30:48.992951 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2" event={"ID":"0f80c25b-98a7-46d0-bf07-26c19f885781","Type":"ContainerStarted","Data":"21a54192dcde69365511cc0c6cbc2b997e44d86061fb138b89cb8c240a849bdb"} Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.406607 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2"] Oct 01 16:30:49 crc kubenswrapper[4726]: E1001 16:30:49.412579 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de9f82a-fe6f-47a8-be2a-0a513407b081" containerName="nmstate-operator" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.412610 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de9f82a-fe6f-47a8-be2a-0a513407b081" containerName="nmstate-operator" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.413798 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5de9f82a-fe6f-47a8-be2a-0a513407b081" containerName="nmstate-operator" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.415321 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.424376 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2"] Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.538290 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-webhook-cert\") pod \"metallb-operator-controller-manager-5d5479c8f-jc5f2\" (UID: \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\") " pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.538371 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82bdc\" (UniqueName: \"kubernetes.io/projected/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-kube-api-access-82bdc\") pod \"metallb-operator-controller-manager-5d5479c8f-jc5f2\" (UID: \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\") " pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.538425 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-apiservice-cert\") pod \"metallb-operator-controller-manager-5d5479c8f-jc5f2\" (UID: \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\") " pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.640496 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-apiservice-cert\") pod \"metallb-operator-controller-manager-5d5479c8f-jc5f2\" (UID: \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\") " pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.640663 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-webhook-cert\") pod \"metallb-operator-controller-manager-5d5479c8f-jc5f2\" (UID: \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\") " pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.640753 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82bdc\" (UniqueName: \"kubernetes.io/projected/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-kube-api-access-82bdc\") pod \"metallb-operator-controller-manager-5d5479c8f-jc5f2\" (UID: \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\") " pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.652079 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-apiservice-cert\") pod \"metallb-operator-controller-manager-5d5479c8f-jc5f2\" (UID: \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\") " pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.652506 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-webhook-cert\") pod \"metallb-operator-controller-manager-5d5479c8f-jc5f2\" (UID: \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\") " pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.673205 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82bdc\" (UniqueName: \"kubernetes.io/projected/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-kube-api-access-82bdc\") pod \"metallb-operator-controller-manager-5d5479c8f-jc5f2\" (UID: \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\") " pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.748981 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.824884 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c168234-66ba-492c-986c-835301bd0275" path="/var/lib/kubelet/pods/4c168234-66ba-492c-986c-835301bd0275/volumes" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.825714 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5de9f82a-fe6f-47a8-be2a-0a513407b081" path="/var/lib/kubelet/pods/5de9f82a-fe6f-47a8-be2a-0a513407b081/volumes" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.826287 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f179fe18-0a34-43df-bb39-dab69a9ee29b" path="/var/lib/kubelet/pods/f179fe18-0a34-43df-bb39-dab69a9ee29b/volumes" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.942366 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l"] Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.943947 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:30:49 crc kubenswrapper[4726]: I1001 16:30:49.986464 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l"] Oct 01 16:30:50 crc kubenswrapper[4726]: I1001 16:30:50.055304 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-webhook-cert\") pod \"metallb-operator-webhook-server-55b5c94945-7tz8l\" (UID: \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\") " pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:30:50 crc kubenswrapper[4726]: I1001 16:30:50.055371 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvv8b\" (UniqueName: \"kubernetes.io/projected/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-kube-api-access-qvv8b\") pod \"metallb-operator-webhook-server-55b5c94945-7tz8l\" (UID: \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\") " pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:30:50 crc kubenswrapper[4726]: I1001 16:30:50.055433 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-apiservice-cert\") pod \"metallb-operator-webhook-server-55b5c94945-7tz8l\" (UID: \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\") " pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:30:50 crc kubenswrapper[4726]: I1001 16:30:50.157405 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-webhook-cert\") pod \"metallb-operator-webhook-server-55b5c94945-7tz8l\" (UID: \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\") " pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:30:50 crc kubenswrapper[4726]: I1001 16:30:50.157852 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvv8b\" (UniqueName: \"kubernetes.io/projected/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-kube-api-access-qvv8b\") pod \"metallb-operator-webhook-server-55b5c94945-7tz8l\" (UID: \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\") " pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:30:50 crc kubenswrapper[4726]: I1001 16:30:50.158020 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-apiservice-cert\") pod \"metallb-operator-webhook-server-55b5c94945-7tz8l\" (UID: \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\") " pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:30:50 crc kubenswrapper[4726]: I1001 16:30:50.163533 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-apiservice-cert\") pod \"metallb-operator-webhook-server-55b5c94945-7tz8l\" (UID: \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\") " pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:30:50 crc kubenswrapper[4726]: I1001 16:30:50.163644 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-webhook-cert\") pod \"metallb-operator-webhook-server-55b5c94945-7tz8l\" (UID: \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\") " pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:30:50 crc kubenswrapper[4726]: I1001 16:30:50.174686 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvv8b\" (UniqueName: \"kubernetes.io/projected/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-kube-api-access-qvv8b\") pod \"metallb-operator-webhook-server-55b5c94945-7tz8l\" (UID: \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\") " pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:30:50 crc kubenswrapper[4726]: I1001 16:30:50.278329 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:30:52 crc kubenswrapper[4726]: I1001 16:30:52.271782 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2"] Oct 01 16:30:52 crc kubenswrapper[4726]: W1001 16:30:52.297262 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6b4559c_bbf0_4f5b_a3e6_d012136cf669.slice/crio-39cad36fc9aa565fdbba64217ea21ee817e1fbc58b3533e407d45353c26e3ee9 WatchSource:0}: Error finding container 39cad36fc9aa565fdbba64217ea21ee817e1fbc58b3533e407d45353c26e3ee9: Status 404 returned error can't find the container with id 39cad36fc9aa565fdbba64217ea21ee817e1fbc58b3533e407d45353c26e3ee9 Oct 01 16:30:52 crc kubenswrapper[4726]: I1001 16:30:52.415404 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l"] Oct 01 16:30:52 crc kubenswrapper[4726]: W1001 16:30:52.423803 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46a1d32f_b37b_48cb_880f_6cbb959a7b3a.slice/crio-cb91a0d7b6a976b9b6bb4205b695ae951e42c9b8bb97fb5f53ee259f64fa6ee0 WatchSource:0}: Error finding container cb91a0d7b6a976b9b6bb4205b695ae951e42c9b8bb97fb5f53ee259f64fa6ee0: Status 404 returned error can't find the container with id cb91a0d7b6a976b9b6bb4205b695ae951e42c9b8bb97fb5f53ee259f64fa6ee0 Oct 01 16:30:53 crc kubenswrapper[4726]: I1001 16:30:53.047111 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xxpvz" event={"ID":"d24b198c-db7b-4057-bed1-6f32d0bac5d3","Type":"ContainerStarted","Data":"d59a7a5977b7316d548cf3d167d5645cd9cebd5e8648b7028f5a032720c903df"} Oct 01 16:30:53 crc kubenswrapper[4726]: I1001 16:30:53.050974 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2" event={"ID":"0f80c25b-98a7-46d0-bf07-26c19f885781","Type":"ContainerStarted","Data":"5e084c34de863116c08bd63a8d9061942c5ad0ead070e76654ce728f1ea5457e"} Oct 01 16:30:53 crc kubenswrapper[4726]: I1001 16:30:53.051141 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2" Oct 01 16:30:53 crc kubenswrapper[4726]: I1001 16:30:53.053345 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" event={"ID":"46a1d32f-b37b-48cb-880f-6cbb959a7b3a","Type":"ContainerStarted","Data":"cb91a0d7b6a976b9b6bb4205b695ae951e42c9b8bb97fb5f53ee259f64fa6ee0"} Oct 01 16:30:53 crc kubenswrapper[4726]: I1001 16:30:53.062338 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-kzwz4" event={"ID":"fc59e6f9-76eb-49fb-9028-9e64820de406","Type":"ContainerStarted","Data":"ff7b45f931f4549649ac65ddeb500bd9111b6125e6ea927b152995d46c53153f"} Oct 01 16:30:53 crc kubenswrapper[4726]: I1001 16:30:53.070394 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm" event={"ID":"5fb896f2-8fbc-40d7-adf4-2a93089171ce","Type":"ContainerStarted","Data":"c47c77a9354bac9308b584ee7387610fc9617bf8e1cc98fbe85558833999bf0d"} Oct 01 16:30:53 crc kubenswrapper[4726]: I1001 16:30:53.072218 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" event={"ID":"f6b4559c-bbf0-4f5b-a3e6-d012136cf669","Type":"ContainerStarted","Data":"39cad36fc9aa565fdbba64217ea21ee817e1fbc58b3533e407d45353c26e3ee9"} Oct 01 16:30:53 crc kubenswrapper[4726]: I1001 16:30:53.077513 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2" podStartSLOduration=2.42885078 podStartE2EDuration="6.077495633s" podCreationTimestamp="2025-10-01 16:30:47 +0000 UTC" firstStartedPulling="2025-10-01 16:30:48.195152858 +0000 UTC m=+1901.096705435" lastFinishedPulling="2025-10-01 16:30:51.843797691 +0000 UTC m=+1904.745350288" observedRunningTime="2025-10-01 16:30:53.069932634 +0000 UTC m=+1905.971485241" watchObservedRunningTime="2025-10-01 16:30:53.077495633 +0000 UTC m=+1905.979048210" Oct 01 16:30:53 crc kubenswrapper[4726]: I1001 16:30:53.126030 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-kzwz4" podStartSLOduration=2.407086434 podStartE2EDuration="7.12601084s" podCreationTimestamp="2025-10-01 16:30:46 +0000 UTC" firstStartedPulling="2025-10-01 16:30:47.177200113 +0000 UTC m=+1900.078752690" lastFinishedPulling="2025-10-01 16:30:51.896124519 +0000 UTC m=+1904.797677096" observedRunningTime="2025-10-01 16:30:53.089515712 +0000 UTC m=+1905.991068289" watchObservedRunningTime="2025-10-01 16:30:53.12601084 +0000 UTC m=+1906.027563417" Oct 01 16:30:53 crc kubenswrapper[4726]: I1001 16:30:53.151009 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vqzfm" podStartSLOduration=3.063262707 podStartE2EDuration="7.150987515s" podCreationTimestamp="2025-10-01 16:30:46 +0000 UTC" firstStartedPulling="2025-10-01 16:30:47.675807005 +0000 UTC m=+1900.577359582" lastFinishedPulling="2025-10-01 16:30:51.763531813 +0000 UTC m=+1904.665084390" observedRunningTime="2025-10-01 16:30:53.113064165 +0000 UTC m=+1906.014616762" watchObservedRunningTime="2025-10-01 16:30:53.150987515 +0000 UTC m=+1906.052540092" Oct 01 16:30:53 crc kubenswrapper[4726]: I1001 16:30:53.163122 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh"] Oct 01 16:30:53 crc kubenswrapper[4726]: I1001 16:30:53.163519 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" podUID="c725f6ed-2c36-43dc-86f4-3a535030832b" containerName="nmstate-console-plugin" containerID="cri-o://9702f03a2d467ad4bbc71677132d4711df89659823db12868d1296226c3c611f" gracePeriod=30 Oct 01 16:30:54 crc kubenswrapper[4726]: I1001 16:30:54.092123 4726 generic.go:334] "Generic (PLEG): container finished" podID="c725f6ed-2c36-43dc-86f4-3a535030832b" containerID="9702f03a2d467ad4bbc71677132d4711df89659823db12868d1296226c3c611f" exitCode=0 Oct 01 16:30:54 crc kubenswrapper[4726]: I1001 16:30:54.092266 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" event={"ID":"c725f6ed-2c36-43dc-86f4-3a535030832b","Type":"ContainerDied","Data":"9702f03a2d467ad4bbc71677132d4711df89659823db12868d1296226c3c611f"} Oct 01 16:30:54 crc kubenswrapper[4726]: I1001 16:30:54.092907 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:30:54 crc kubenswrapper[4726]: I1001 16:30:54.232008 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" Oct 01 16:30:54 crc kubenswrapper[4726]: I1001 16:30:54.363780 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c725f6ed-2c36-43dc-86f4-3a535030832b-nginx-conf\") pod \"c725f6ed-2c36-43dc-86f4-3a535030832b\" (UID: \"c725f6ed-2c36-43dc-86f4-3a535030832b\") " Oct 01 16:30:54 crc kubenswrapper[4726]: I1001 16:30:54.363825 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9rcr\" (UniqueName: \"kubernetes.io/projected/c725f6ed-2c36-43dc-86f4-3a535030832b-kube-api-access-d9rcr\") pod \"c725f6ed-2c36-43dc-86f4-3a535030832b\" (UID: \"c725f6ed-2c36-43dc-86f4-3a535030832b\") " Oct 01 16:30:54 crc kubenswrapper[4726]: I1001 16:30:54.363886 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c725f6ed-2c36-43dc-86f4-3a535030832b-plugin-serving-cert\") pod \"c725f6ed-2c36-43dc-86f4-3a535030832b\" (UID: \"c725f6ed-2c36-43dc-86f4-3a535030832b\") " Oct 01 16:30:54 crc kubenswrapper[4726]: I1001 16:30:54.387781 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c725f6ed-2c36-43dc-86f4-3a535030832b-plugin-serving-cert" (OuterVolumeSpecName: "plugin-serving-cert") pod "c725f6ed-2c36-43dc-86f4-3a535030832b" (UID: "c725f6ed-2c36-43dc-86f4-3a535030832b"). InnerVolumeSpecName "plugin-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:30:54 crc kubenswrapper[4726]: I1001 16:30:54.387861 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c725f6ed-2c36-43dc-86f4-3a535030832b-kube-api-access-d9rcr" (OuterVolumeSpecName: "kube-api-access-d9rcr") pod "c725f6ed-2c36-43dc-86f4-3a535030832b" (UID: "c725f6ed-2c36-43dc-86f4-3a535030832b"). InnerVolumeSpecName "kube-api-access-d9rcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:30:54 crc kubenswrapper[4726]: I1001 16:30:54.408340 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c725f6ed-2c36-43dc-86f4-3a535030832b-nginx-conf" (OuterVolumeSpecName: "nginx-conf") pod "c725f6ed-2c36-43dc-86f4-3a535030832b" (UID: "c725f6ed-2c36-43dc-86f4-3a535030832b"). InnerVolumeSpecName "nginx-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:30:54 crc kubenswrapper[4726]: I1001 16:30:54.465920 4726 reconciler_common.go:293] "Volume detached for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c725f6ed-2c36-43dc-86f4-3a535030832b-nginx-conf\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:54 crc kubenswrapper[4726]: I1001 16:30:54.465964 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9rcr\" (UniqueName: \"kubernetes.io/projected/c725f6ed-2c36-43dc-86f4-3a535030832b-kube-api-access-d9rcr\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:54 crc kubenswrapper[4726]: I1001 16:30:54.465978 4726 reconciler_common.go:293] "Volume detached for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c725f6ed-2c36-43dc-86f4-3a535030832b-plugin-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:55 crc kubenswrapper[4726]: I1001 16:30:55.108064 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" event={"ID":"c725f6ed-2c36-43dc-86f4-3a535030832b","Type":"ContainerDied","Data":"3140a050a2ca6affa03455f3cde5dcfbf5d065bce40d938018a23712119be6cf"} Oct 01 16:30:55 crc kubenswrapper[4726]: I1001 16:30:55.108101 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh" Oct 01 16:30:55 crc kubenswrapper[4726]: I1001 16:30:55.108125 4726 scope.go:117] "RemoveContainer" containerID="9702f03a2d467ad4bbc71677132d4711df89659823db12868d1296226c3c611f" Oct 01 16:30:55 crc kubenswrapper[4726]: I1001 16:30:55.214196 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh"] Oct 01 16:30:55 crc kubenswrapper[4726]: I1001 16:30:55.224443 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-xffxh"] Oct 01 16:30:55 crc kubenswrapper[4726]: I1001 16:30:55.827088 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c725f6ed-2c36-43dc-86f4-3a535030832b" path="/var/lib/kubelet/pods/c725f6ed-2c36-43dc-86f4-3a535030832b/volumes" Oct 01 16:30:57 crc kubenswrapper[4726]: I1001 16:30:57.163600 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-kzwz4" Oct 01 16:31:03 crc kubenswrapper[4726]: I1001 16:31:03.200557 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" event={"ID":"f6b4559c-bbf0-4f5b-a3e6-d012136cf669","Type":"ContainerStarted","Data":"afa4ef815c35f4037fe2a0de09c92c79122752989b8ce5704f7f3e8878d5ed08"} Oct 01 16:31:03 crc kubenswrapper[4726]: I1001 16:31:03.201094 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:31:03 crc kubenswrapper[4726]: I1001 16:31:03.204151 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xxpvz" event={"ID":"d24b198c-db7b-4057-bed1-6f32d0bac5d3","Type":"ContainerStarted","Data":"ebafc8706fcda5df836aed35b09d997751a99fc8eedf103817b50978016fb8f3"} Oct 01 16:31:03 crc kubenswrapper[4726]: I1001 16:31:03.207513 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" event={"ID":"46a1d32f-b37b-48cb-880f-6cbb959a7b3a","Type":"ContainerStarted","Data":"6b34947dbaee53f404e0aa6070a1f595e64dc9e43284e019093d0a9634b1048f"} Oct 01 16:31:03 crc kubenswrapper[4726]: I1001 16:31:03.207704 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:31:03 crc kubenswrapper[4726]: I1001 16:31:03.225763 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" podStartSLOduration=4.189354147 podStartE2EDuration="14.225742299s" podCreationTimestamp="2025-10-01 16:30:49 +0000 UTC" firstStartedPulling="2025-10-01 16:30:52.308703955 +0000 UTC m=+1905.210256532" lastFinishedPulling="2025-10-01 16:31:02.345092107 +0000 UTC m=+1915.246644684" observedRunningTime="2025-10-01 16:31:03.223550366 +0000 UTC m=+1916.125102963" watchObservedRunningTime="2025-10-01 16:31:03.225742299 +0000 UTC m=+1916.127294876" Oct 01 16:31:03 crc kubenswrapper[4726]: I1001 16:31:03.258929 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" podStartSLOduration=4.420223564 podStartE2EDuration="14.258891711s" podCreationTimestamp="2025-10-01 16:30:49 +0000 UTC" firstStartedPulling="2025-10-01 16:30:52.427815 +0000 UTC m=+1905.329367577" lastFinishedPulling="2025-10-01 16:31:02.266483147 +0000 UTC m=+1915.168035724" observedRunningTime="2025-10-01 16:31:03.255338888 +0000 UTC m=+1916.156891475" watchObservedRunningTime="2025-10-01 16:31:03.258891711 +0000 UTC m=+1916.160444288" Oct 01 16:31:03 crc kubenswrapper[4726]: I1001 16:31:03.283029 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xxpvz" podStartSLOduration=2.404980807 podStartE2EDuration="16.28300776s" podCreationTimestamp="2025-10-01 16:30:47 +0000 UTC" firstStartedPulling="2025-10-01 16:30:48.318780093 +0000 UTC m=+1901.220332670" lastFinishedPulling="2025-10-01 16:31:02.196807046 +0000 UTC m=+1915.098359623" observedRunningTime="2025-10-01 16:31:03.274371 +0000 UTC m=+1916.175923577" watchObservedRunningTime="2025-10-01 16:31:03.28300776 +0000 UTC m=+1916.184560337" Oct 01 16:31:07 crc kubenswrapper[4726]: I1001 16:31:07.672957 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-qrkz2" Oct 01 16:31:11 crc kubenswrapper[4726]: I1001 16:31:11.281538 4726 generic.go:334] "Generic (PLEG): container finished" podID="b8c389e9-bf2c-4e85-a2ec-fc334262347f" containerID="11923a927cd267b9bde7ef93092ef6331c479aaae8fc9f8519e757a6f5ff33c6" exitCode=0 Oct 01 16:31:11 crc kubenswrapper[4726]: I1001 16:31:11.282019 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" event={"ID":"b8c389e9-bf2c-4e85-a2ec-fc334262347f","Type":"ContainerDied","Data":"11923a927cd267b9bde7ef93092ef6331c479aaae8fc9f8519e757a6f5ff33c6"} Oct 01 16:31:12 crc kubenswrapper[4726]: I1001 16:31:12.898327 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" Oct 01 16:31:12 crc kubenswrapper[4726]: I1001 16:31:12.985651 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c389e9-bf2c-4e85-a2ec-fc334262347f-inventory\") pod \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\" (UID: \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\") " Oct 01 16:31:12 crc kubenswrapper[4726]: I1001 16:31:12.985932 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8c389e9-bf2c-4e85-a2ec-fc334262347f-ssh-key\") pod \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\" (UID: \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\") " Oct 01 16:31:12 crc kubenswrapper[4726]: I1001 16:31:12.985998 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db4v8\" (UniqueName: \"kubernetes.io/projected/b8c389e9-bf2c-4e85-a2ec-fc334262347f-kube-api-access-db4v8\") pod \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\" (UID: \"b8c389e9-bf2c-4e85-a2ec-fc334262347f\") " Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.003416 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8c389e9-bf2c-4e85-a2ec-fc334262347f-kube-api-access-db4v8" (OuterVolumeSpecName: "kube-api-access-db4v8") pod "b8c389e9-bf2c-4e85-a2ec-fc334262347f" (UID: "b8c389e9-bf2c-4e85-a2ec-fc334262347f"). InnerVolumeSpecName "kube-api-access-db4v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.023289 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8c389e9-bf2c-4e85-a2ec-fc334262347f-inventory" (OuterVolumeSpecName: "inventory") pod "b8c389e9-bf2c-4e85-a2ec-fc334262347f" (UID: "b8c389e9-bf2c-4e85-a2ec-fc334262347f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.041579 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8c389e9-bf2c-4e85-a2ec-fc334262347f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b8c389e9-bf2c-4e85-a2ec-fc334262347f" (UID: "b8c389e9-bf2c-4e85-a2ec-fc334262347f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.089469 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c389e9-bf2c-4e85-a2ec-fc334262347f-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.089808 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8c389e9-bf2c-4e85-a2ec-fc334262347f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.089824 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db4v8\" (UniqueName: \"kubernetes.io/projected/b8c389e9-bf2c-4e85-a2ec-fc334262347f-kube-api-access-db4v8\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.300322 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" event={"ID":"b8c389e9-bf2c-4e85-a2ec-fc334262347f","Type":"ContainerDied","Data":"3a19c533e4166de6016702ab42c88d76acdf82434b7fddb06cb3d00efa62e961"} Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.300362 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a19c533e4166de6016702ab42c88d76acdf82434b7fddb06cb3d00efa62e961" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.300415 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4s76w" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.401749 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mpm5k"] Oct 01 16:31:13 crc kubenswrapper[4726]: E1001 16:31:13.402165 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8c389e9-bf2c-4e85-a2ec-fc334262347f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.402183 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8c389e9-bf2c-4e85-a2ec-fc334262347f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:31:13 crc kubenswrapper[4726]: E1001 16:31:13.402194 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c725f6ed-2c36-43dc-86f4-3a535030832b" containerName="nmstate-console-plugin" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.402202 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c725f6ed-2c36-43dc-86f4-3a535030832b" containerName="nmstate-console-plugin" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.402406 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8c389e9-bf2c-4e85-a2ec-fc334262347f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.402439 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c725f6ed-2c36-43dc-86f4-3a535030832b" containerName="nmstate-console-plugin" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.403765 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.407144 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.407487 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.411249 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.412487 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.416756 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mpm5k"] Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.502325 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e03d54c6-e53e-4837-9d0d-a9183775699e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mpm5k\" (UID: \"e03d54c6-e53e-4837-9d0d-a9183775699e\") " pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.502658 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e03d54c6-e53e-4837-9d0d-a9183775699e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mpm5k\" (UID: \"e03d54c6-e53e-4837-9d0d-a9183775699e\") " pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.502681 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcmw8\" (UniqueName: \"kubernetes.io/projected/e03d54c6-e53e-4837-9d0d-a9183775699e-kube-api-access-dcmw8\") pod \"ssh-known-hosts-edpm-deployment-mpm5k\" (UID: \"e03d54c6-e53e-4837-9d0d-a9183775699e\") " pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.604621 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e03d54c6-e53e-4837-9d0d-a9183775699e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mpm5k\" (UID: \"e03d54c6-e53e-4837-9d0d-a9183775699e\") " pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.604729 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e03d54c6-e53e-4837-9d0d-a9183775699e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mpm5k\" (UID: \"e03d54c6-e53e-4837-9d0d-a9183775699e\") " pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.604756 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcmw8\" (UniqueName: \"kubernetes.io/projected/e03d54c6-e53e-4837-9d0d-a9183775699e-kube-api-access-dcmw8\") pod \"ssh-known-hosts-edpm-deployment-mpm5k\" (UID: \"e03d54c6-e53e-4837-9d0d-a9183775699e\") " pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.611332 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e03d54c6-e53e-4837-9d0d-a9183775699e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mpm5k\" (UID: \"e03d54c6-e53e-4837-9d0d-a9183775699e\") " pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.612353 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e03d54c6-e53e-4837-9d0d-a9183775699e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mpm5k\" (UID: \"e03d54c6-e53e-4837-9d0d-a9183775699e\") " pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.625567 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcmw8\" (UniqueName: \"kubernetes.io/projected/e03d54c6-e53e-4837-9d0d-a9183775699e-kube-api-access-dcmw8\") pod \"ssh-known-hosts-edpm-deployment-mpm5k\" (UID: \"e03d54c6-e53e-4837-9d0d-a9183775699e\") " pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" Oct 01 16:31:13 crc kubenswrapper[4726]: I1001 16:31:13.729487 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" Oct 01 16:31:14 crc kubenswrapper[4726]: I1001 16:31:14.494770 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mpm5k"] Oct 01 16:31:15 crc kubenswrapper[4726]: I1001 16:31:15.327432 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" event={"ID":"e03d54c6-e53e-4837-9d0d-a9183775699e","Type":"ContainerStarted","Data":"188e4eaa5dd33fed8c6ec2066f2d1922024b41035d68b4e7bb2174ecc3520ca8"} Oct 01 16:31:16 crc kubenswrapper[4726]: I1001 16:31:16.338865 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" event={"ID":"e03d54c6-e53e-4837-9d0d-a9183775699e","Type":"ContainerStarted","Data":"107288a437d6b572318ae40fea272c426b937a7bee108f8abd026036ad8c9851"} Oct 01 16:31:16 crc kubenswrapper[4726]: I1001 16:31:16.365745 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" podStartSLOduration=2.851554872 podStartE2EDuration="3.365717425s" podCreationTimestamp="2025-10-01 16:31:13 +0000 UTC" firstStartedPulling="2025-10-01 16:31:14.512242168 +0000 UTC m=+1927.413794745" lastFinishedPulling="2025-10-01 16:31:15.026404721 +0000 UTC m=+1927.927957298" observedRunningTime="2025-10-01 16:31:16.365595041 +0000 UTC m=+1929.267147638" watchObservedRunningTime="2025-10-01 16:31:16.365717425 +0000 UTC m=+1929.267270022" Oct 01 16:31:20 crc kubenswrapper[4726]: I1001 16:31:20.287241 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:31:20 crc kubenswrapper[4726]: I1001 16:31:20.373631 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48"] Oct 01 16:31:20 crc kubenswrapper[4726]: I1001 16:31:20.374003 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" podUID="daa5265d-09f5-4a83-96a0-415448d42bcd" containerName="webhook-server" containerID="cri-o://91ef4d875b12d908c04b4179a391a99b64695c29ccb91c2cf6cfdb806b9e00b0" gracePeriod=2 Oct 01 16:31:20 crc kubenswrapper[4726]: I1001 16:31:20.396196 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48"] Oct 01 16:31:20 crc kubenswrapper[4726]: I1001 16:31:20.922506 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.112118 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/daa5265d-09f5-4a83-96a0-415448d42bcd-webhook-cert\") pod \"daa5265d-09f5-4a83-96a0-415448d42bcd\" (UID: \"daa5265d-09f5-4a83-96a0-415448d42bcd\") " Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.112500 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j66rg\" (UniqueName: \"kubernetes.io/projected/daa5265d-09f5-4a83-96a0-415448d42bcd-kube-api-access-j66rg\") pod \"daa5265d-09f5-4a83-96a0-415448d42bcd\" (UID: \"daa5265d-09f5-4a83-96a0-415448d42bcd\") " Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.112589 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/daa5265d-09f5-4a83-96a0-415448d42bcd-apiservice-cert\") pod \"daa5265d-09f5-4a83-96a0-415448d42bcd\" (UID: \"daa5265d-09f5-4a83-96a0-415448d42bcd\") " Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.124968 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daa5265d-09f5-4a83-96a0-415448d42bcd-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "daa5265d-09f5-4a83-96a0-415448d42bcd" (UID: "daa5265d-09f5-4a83-96a0-415448d42bcd"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.125082 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daa5265d-09f5-4a83-96a0-415448d42bcd-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "daa5265d-09f5-4a83-96a0-415448d42bcd" (UID: "daa5265d-09f5-4a83-96a0-415448d42bcd"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.133505 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daa5265d-09f5-4a83-96a0-415448d42bcd-kube-api-access-j66rg" (OuterVolumeSpecName: "kube-api-access-j66rg") pod "daa5265d-09f5-4a83-96a0-415448d42bcd" (UID: "daa5265d-09f5-4a83-96a0-415448d42bcd"). InnerVolumeSpecName "kube-api-access-j66rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.215700 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j66rg\" (UniqueName: \"kubernetes.io/projected/daa5265d-09f5-4a83-96a0-415448d42bcd-kube-api-access-j66rg\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.215741 4726 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/daa5265d-09f5-4a83-96a0-415448d42bcd-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.215763 4726 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/daa5265d-09f5-4a83-96a0-415448d42bcd-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.453470 4726 generic.go:334] "Generic (PLEG): container finished" podID="daa5265d-09f5-4a83-96a0-415448d42bcd" containerID="91ef4d875b12d908c04b4179a391a99b64695c29ccb91c2cf6cfdb806b9e00b0" exitCode=0 Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.453570 4726 scope.go:117] "RemoveContainer" containerID="91ef4d875b12d908c04b4179a391a99b64695c29ccb91c2cf6cfdb806b9e00b0" Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.453781 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d69b94c67-gnx48" Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.489795 4726 scope.go:117] "RemoveContainer" containerID="91ef4d875b12d908c04b4179a391a99b64695c29ccb91c2cf6cfdb806b9e00b0" Oct 01 16:31:21 crc kubenswrapper[4726]: E1001 16:31:21.490987 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91ef4d875b12d908c04b4179a391a99b64695c29ccb91c2cf6cfdb806b9e00b0\": container with ID starting with 91ef4d875b12d908c04b4179a391a99b64695c29ccb91c2cf6cfdb806b9e00b0 not found: ID does not exist" containerID="91ef4d875b12d908c04b4179a391a99b64695c29ccb91c2cf6cfdb806b9e00b0" Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.491135 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91ef4d875b12d908c04b4179a391a99b64695c29ccb91c2cf6cfdb806b9e00b0"} err="failed to get container status \"91ef4d875b12d908c04b4179a391a99b64695c29ccb91c2cf6cfdb806b9e00b0\": rpc error: code = NotFound desc = could not find container \"91ef4d875b12d908c04b4179a391a99b64695c29ccb91c2cf6cfdb806b9e00b0\": container with ID starting with 91ef4d875b12d908c04b4179a391a99b64695c29ccb91c2cf6cfdb806b9e00b0 not found: ID does not exist" Oct 01 16:31:21 crc kubenswrapper[4726]: I1001 16:31:21.824965 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daa5265d-09f5-4a83-96a0-415448d42bcd" path="/var/lib/kubelet/pods/daa5265d-09f5-4a83-96a0-415448d42bcd/volumes" Oct 01 16:31:23 crc kubenswrapper[4726]: I1001 16:31:23.506816 4726 generic.go:334] "Generic (PLEG): container finished" podID="e03d54c6-e53e-4837-9d0d-a9183775699e" containerID="107288a437d6b572318ae40fea272c426b937a7bee108f8abd026036ad8c9851" exitCode=0 Oct 01 16:31:23 crc kubenswrapper[4726]: I1001 16:31:23.506861 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" event={"ID":"e03d54c6-e53e-4837-9d0d-a9183775699e","Type":"ContainerDied","Data":"107288a437d6b572318ae40fea272c426b937a7bee108f8abd026036ad8c9851"} Oct 01 16:31:23 crc kubenswrapper[4726]: I1001 16:31:23.540804 4726 scope.go:117] "RemoveContainer" containerID="630c37bbb33455a45b1327918f394cbce39071d19b9ee391447795f0d1909969" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.035386 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.053858 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e03d54c6-e53e-4837-9d0d-a9183775699e-ssh-key-openstack-edpm-ipam\") pod \"e03d54c6-e53e-4837-9d0d-a9183775699e\" (UID: \"e03d54c6-e53e-4837-9d0d-a9183775699e\") " Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.054074 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcmw8\" (UniqueName: \"kubernetes.io/projected/e03d54c6-e53e-4837-9d0d-a9183775699e-kube-api-access-dcmw8\") pod \"e03d54c6-e53e-4837-9d0d-a9183775699e\" (UID: \"e03d54c6-e53e-4837-9d0d-a9183775699e\") " Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.054115 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e03d54c6-e53e-4837-9d0d-a9183775699e-inventory-0\") pod \"e03d54c6-e53e-4837-9d0d-a9183775699e\" (UID: \"e03d54c6-e53e-4837-9d0d-a9183775699e\") " Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.069361 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e03d54c6-e53e-4837-9d0d-a9183775699e-kube-api-access-dcmw8" (OuterVolumeSpecName: "kube-api-access-dcmw8") pod "e03d54c6-e53e-4837-9d0d-a9183775699e" (UID: "e03d54c6-e53e-4837-9d0d-a9183775699e"). InnerVolumeSpecName "kube-api-access-dcmw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.095539 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e03d54c6-e53e-4837-9d0d-a9183775699e-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "e03d54c6-e53e-4837-9d0d-a9183775699e" (UID: "e03d54c6-e53e-4837-9d0d-a9183775699e"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.116276 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e03d54c6-e53e-4837-9d0d-a9183775699e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e03d54c6-e53e-4837-9d0d-a9183775699e" (UID: "e03d54c6-e53e-4837-9d0d-a9183775699e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.162432 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e03d54c6-e53e-4837-9d0d-a9183775699e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.162490 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcmw8\" (UniqueName: \"kubernetes.io/projected/e03d54c6-e53e-4837-9d0d-a9183775699e-kube-api-access-dcmw8\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.162510 4726 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e03d54c6-e53e-4837-9d0d-a9183775699e-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.527854 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" event={"ID":"e03d54c6-e53e-4837-9d0d-a9183775699e","Type":"ContainerDied","Data":"188e4eaa5dd33fed8c6ec2066f2d1922024b41035d68b4e7bb2174ecc3520ca8"} Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.528153 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="188e4eaa5dd33fed8c6ec2066f2d1922024b41035d68b4e7bb2174ecc3520ca8" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.527921 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mpm5k" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.623893 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b"] Oct 01 16:31:25 crc kubenswrapper[4726]: E1001 16:31:25.624551 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daa5265d-09f5-4a83-96a0-415448d42bcd" containerName="webhook-server" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.624619 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="daa5265d-09f5-4a83-96a0-415448d42bcd" containerName="webhook-server" Oct 01 16:31:25 crc kubenswrapper[4726]: E1001 16:31:25.624730 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e03d54c6-e53e-4837-9d0d-a9183775699e" containerName="ssh-known-hosts-edpm-deployment" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.624811 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e03d54c6-e53e-4837-9d0d-a9183775699e" containerName="ssh-known-hosts-edpm-deployment" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.625101 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e03d54c6-e53e-4837-9d0d-a9183775699e" containerName="ssh-known-hosts-edpm-deployment" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.625187 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="daa5265d-09f5-4a83-96a0-415448d42bcd" containerName="webhook-server" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.625872 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.627724 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.627931 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.628907 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.628902 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.635907 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b"] Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.674621 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdw7n\" (UniqueName: \"kubernetes.io/projected/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-kube-api-access-hdw7n\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6fb8b\" (UID: \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.674907 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6fb8b\" (UID: \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.675238 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6fb8b\" (UID: \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.777788 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6fb8b\" (UID: \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.777924 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6fb8b\" (UID: \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.778042 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdw7n\" (UniqueName: \"kubernetes.io/projected/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-kube-api-access-hdw7n\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6fb8b\" (UID: \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.784085 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6fb8b\" (UID: \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.784596 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6fb8b\" (UID: \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.804889 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdw7n\" (UniqueName: \"kubernetes.io/projected/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-kube-api-access-hdw7n\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6fb8b\" (UID: \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" Oct 01 16:31:25 crc kubenswrapper[4726]: I1001 16:31:25.943679 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" Oct 01 16:31:26 crc kubenswrapper[4726]: I1001 16:31:26.483339 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b"] Oct 01 16:31:26 crc kubenswrapper[4726]: W1001 16:31:26.501229 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1537a5c5_290d_4e30_ba28_f2b90d0d1fe3.slice/crio-eb1226303c5f209651c433f30162c7d6966f4c195ea55231397c09d8213c70b5 WatchSource:0}: Error finding container eb1226303c5f209651c433f30162c7d6966f4c195ea55231397c09d8213c70b5: Status 404 returned error can't find the container with id eb1226303c5f209651c433f30162c7d6966f4c195ea55231397c09d8213c70b5 Oct 01 16:31:26 crc kubenswrapper[4726]: I1001 16:31:26.542638 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" event={"ID":"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3","Type":"ContainerStarted","Data":"eb1226303c5f209651c433f30162c7d6966f4c195ea55231397c09d8213c70b5"} Oct 01 16:31:27 crc kubenswrapper[4726]: I1001 16:31:27.551912 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" event={"ID":"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3","Type":"ContainerStarted","Data":"8247c88508c8ae99812c82ea52b1bbbcb2b3aab9616d60db7b2e59943945b5aa"} Oct 01 16:31:36 crc kubenswrapper[4726]: I1001 16:31:36.637086 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" event={"ID":"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3","Type":"ContainerDied","Data":"8247c88508c8ae99812c82ea52b1bbbcb2b3aab9616d60db7b2e59943945b5aa"} Oct 01 16:31:36 crc kubenswrapper[4726]: I1001 16:31:36.637089 4726 generic.go:334] "Generic (PLEG): container finished" podID="1537a5c5-290d-4e30-ba28-f2b90d0d1fe3" containerID="8247c88508c8ae99812c82ea52b1bbbcb2b3aab9616d60db7b2e59943945b5aa" exitCode=0 Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.218866 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.237821 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-ssh-key\") pod \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\" (UID: \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\") " Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.237882 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdw7n\" (UniqueName: \"kubernetes.io/projected/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-kube-api-access-hdw7n\") pod \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\" (UID: \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\") " Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.237907 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-inventory\") pod \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\" (UID: \"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3\") " Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.260268 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-kube-api-access-hdw7n" (OuterVolumeSpecName: "kube-api-access-hdw7n") pod "1537a5c5-290d-4e30-ba28-f2b90d0d1fe3" (UID: "1537a5c5-290d-4e30-ba28-f2b90d0d1fe3"). InnerVolumeSpecName "kube-api-access-hdw7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.268888 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1537a5c5-290d-4e30-ba28-f2b90d0d1fe3" (UID: "1537a5c5-290d-4e30-ba28-f2b90d0d1fe3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.284997 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-inventory" (OuterVolumeSpecName: "inventory") pod "1537a5c5-290d-4e30-ba28-f2b90d0d1fe3" (UID: "1537a5c5-290d-4e30-ba28-f2b90d0d1fe3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.341091 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.341127 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdw7n\" (UniqueName: \"kubernetes.io/projected/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-kube-api-access-hdw7n\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.341154 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1537a5c5-290d-4e30-ba28-f2b90d0d1fe3-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.654468 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" event={"ID":"1537a5c5-290d-4e30-ba28-f2b90d0d1fe3","Type":"ContainerDied","Data":"eb1226303c5f209651c433f30162c7d6966f4c195ea55231397c09d8213c70b5"} Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.654505 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb1226303c5f209651c433f30162c7d6966f4c195ea55231397c09d8213c70b5" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.654515 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6fb8b" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.730996 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c"] Oct 01 16:31:38 crc kubenswrapper[4726]: E1001 16:31:38.732518 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1537a5c5-290d-4e30-ba28-f2b90d0d1fe3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.732561 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1537a5c5-290d-4e30-ba28-f2b90d0d1fe3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.732816 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1537a5c5-290d-4e30-ba28-f2b90d0d1fe3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.733551 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.735281 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.740317 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.740360 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.740558 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.752472 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c"] Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.851722 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/016c122e-b1fa-43a1-a806-93533e24b8f6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c\" (UID: \"016c122e-b1fa-43a1-a806-93533e24b8f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.851820 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/016c122e-b1fa-43a1-a806-93533e24b8f6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c\" (UID: \"016c122e-b1fa-43a1-a806-93533e24b8f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.852621 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9wfb\" (UniqueName: \"kubernetes.io/projected/016c122e-b1fa-43a1-a806-93533e24b8f6-kube-api-access-f9wfb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c\" (UID: \"016c122e-b1fa-43a1-a806-93533e24b8f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.955656 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/016c122e-b1fa-43a1-a806-93533e24b8f6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c\" (UID: \"016c122e-b1fa-43a1-a806-93533e24b8f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.955747 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/016c122e-b1fa-43a1-a806-93533e24b8f6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c\" (UID: \"016c122e-b1fa-43a1-a806-93533e24b8f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.955815 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9wfb\" (UniqueName: \"kubernetes.io/projected/016c122e-b1fa-43a1-a806-93533e24b8f6-kube-api-access-f9wfb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c\" (UID: \"016c122e-b1fa-43a1-a806-93533e24b8f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.967454 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/016c122e-b1fa-43a1-a806-93533e24b8f6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c\" (UID: \"016c122e-b1fa-43a1-a806-93533e24b8f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.967872 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/016c122e-b1fa-43a1-a806-93533e24b8f6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c\" (UID: \"016c122e-b1fa-43a1-a806-93533e24b8f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" Oct 01 16:31:38 crc kubenswrapper[4726]: I1001 16:31:38.982310 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9wfb\" (UniqueName: \"kubernetes.io/projected/016c122e-b1fa-43a1-a806-93533e24b8f6-kube-api-access-f9wfb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c\" (UID: \"016c122e-b1fa-43a1-a806-93533e24b8f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" Oct 01 16:31:39 crc kubenswrapper[4726]: I1001 16:31:39.052529 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" Oct 01 16:31:39 crc kubenswrapper[4726]: I1001 16:31:39.590118 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c"] Oct 01 16:31:39 crc kubenswrapper[4726]: I1001 16:31:39.665683 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" event={"ID":"016c122e-b1fa-43a1-a806-93533e24b8f6","Type":"ContainerStarted","Data":"659f81010c67cb4391374bbbcfb54ad185b52778f800484251b4130635dfe147"} Oct 01 16:31:39 crc kubenswrapper[4726]: I1001 16:31:39.752173 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:31:39 crc kubenswrapper[4726]: I1001 16:31:39.820579 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh"] Oct 01 16:31:39 crc kubenswrapper[4726]: I1001 16:31:39.820758 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" podUID="94e8d60f-8f3a-448a-9096-163826d5ad92" containerName="manager" containerID="cri-o://f63336a280fb49c32c3b420b5e53614ecd913f739bb694a58658e325e6691317" gracePeriod=10 Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.382350 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.481438 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94e8d60f-8f3a-448a-9096-163826d5ad92-apiservice-cert\") pod \"94e8d60f-8f3a-448a-9096-163826d5ad92\" (UID: \"94e8d60f-8f3a-448a-9096-163826d5ad92\") " Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.481482 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94e8d60f-8f3a-448a-9096-163826d5ad92-webhook-cert\") pod \"94e8d60f-8f3a-448a-9096-163826d5ad92\" (UID: \"94e8d60f-8f3a-448a-9096-163826d5ad92\") " Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.481524 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khxpb\" (UniqueName: \"kubernetes.io/projected/94e8d60f-8f3a-448a-9096-163826d5ad92-kube-api-access-khxpb\") pod \"94e8d60f-8f3a-448a-9096-163826d5ad92\" (UID: \"94e8d60f-8f3a-448a-9096-163826d5ad92\") " Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.488379 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94e8d60f-8f3a-448a-9096-163826d5ad92-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "94e8d60f-8f3a-448a-9096-163826d5ad92" (UID: "94e8d60f-8f3a-448a-9096-163826d5ad92"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.488425 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94e8d60f-8f3a-448a-9096-163826d5ad92-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "94e8d60f-8f3a-448a-9096-163826d5ad92" (UID: "94e8d60f-8f3a-448a-9096-163826d5ad92"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.490641 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94e8d60f-8f3a-448a-9096-163826d5ad92-kube-api-access-khxpb" (OuterVolumeSpecName: "kube-api-access-khxpb") pod "94e8d60f-8f3a-448a-9096-163826d5ad92" (UID: "94e8d60f-8f3a-448a-9096-163826d5ad92"). InnerVolumeSpecName "kube-api-access-khxpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.584493 4726 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94e8d60f-8f3a-448a-9096-163826d5ad92-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.584538 4726 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94e8d60f-8f3a-448a-9096-163826d5ad92-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.584555 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khxpb\" (UniqueName: \"kubernetes.io/projected/94e8d60f-8f3a-448a-9096-163826d5ad92-kube-api-access-khxpb\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.678345 4726 generic.go:334] "Generic (PLEG): container finished" podID="94e8d60f-8f3a-448a-9096-163826d5ad92" containerID="f63336a280fb49c32c3b420b5e53614ecd913f739bb694a58658e325e6691317" exitCode=0 Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.678388 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" event={"ID":"94e8d60f-8f3a-448a-9096-163826d5ad92","Type":"ContainerDied","Data":"f63336a280fb49c32c3b420b5e53614ecd913f739bb694a58658e325e6691317"} Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.678649 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" event={"ID":"94e8d60f-8f3a-448a-9096-163826d5ad92","Type":"ContainerDied","Data":"c10990bc3975e695e6ded271396d6eb2fd0ed59d64ec23c6caf3c82be4bdb321"} Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.678686 4726 scope.go:117] "RemoveContainer" containerID="f63336a280fb49c32c3b420b5e53614ecd913f739bb694a58658e325e6691317" Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.678412 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh" Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.710932 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh"] Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.721206 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/metallb-operator-controller-manager-86d5dd6867-8btsh"] Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.724857 4726 scope.go:117] "RemoveContainer" containerID="f63336a280fb49c32c3b420b5e53614ecd913f739bb694a58658e325e6691317" Oct 01 16:31:40 crc kubenswrapper[4726]: E1001 16:31:40.725293 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f63336a280fb49c32c3b420b5e53614ecd913f739bb694a58658e325e6691317\": container with ID starting with f63336a280fb49c32c3b420b5e53614ecd913f739bb694a58658e325e6691317 not found: ID does not exist" containerID="f63336a280fb49c32c3b420b5e53614ecd913f739bb694a58658e325e6691317" Oct 01 16:31:40 crc kubenswrapper[4726]: I1001 16:31:40.725320 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f63336a280fb49c32c3b420b5e53614ecd913f739bb694a58658e325e6691317"} err="failed to get container status \"f63336a280fb49c32c3b420b5e53614ecd913f739bb694a58658e325e6691317\": rpc error: code = NotFound desc = could not find container \"f63336a280fb49c32c3b420b5e53614ecd913f739bb694a58658e325e6691317\": container with ID starting with f63336a280fb49c32c3b420b5e53614ecd913f739bb694a58658e325e6691317 not found: ID does not exist" Oct 01 16:31:41 crc kubenswrapper[4726]: I1001 16:31:41.689781 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" event={"ID":"016c122e-b1fa-43a1-a806-93533e24b8f6","Type":"ContainerStarted","Data":"df3c0d982b9e4d802e28760d1df645b84c6d7054906c1aeb777d85584d9c1b0d"} Oct 01 16:31:41 crc kubenswrapper[4726]: I1001 16:31:41.719454 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" podStartSLOduration=2.919150621 podStartE2EDuration="3.71943043s" podCreationTimestamp="2025-10-01 16:31:38 +0000 UTC" firstStartedPulling="2025-10-01 16:31:39.601156432 +0000 UTC m=+1952.502709009" lastFinishedPulling="2025-10-01 16:31:40.401436241 +0000 UTC m=+1953.302988818" observedRunningTime="2025-10-01 16:31:41.707817933 +0000 UTC m=+1954.609370500" watchObservedRunningTime="2025-10-01 16:31:41.71943043 +0000 UTC m=+1954.620983007" Oct 01 16:31:41 crc kubenswrapper[4726]: I1001 16:31:41.822374 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94e8d60f-8f3a-448a-9096-163826d5ad92" path="/var/lib/kubelet/pods/94e8d60f-8f3a-448a-9096-163826d5ad92/volumes" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.667461 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr"] Oct 01 16:31:48 crc kubenswrapper[4726]: E1001 16:31:48.668355 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94e8d60f-8f3a-448a-9096-163826d5ad92" containerName="manager" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.668370 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="94e8d60f-8f3a-448a-9096-163826d5ad92" containerName="manager" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.668632 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="94e8d60f-8f3a-448a-9096-163826d5ad92" containerName="manager" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.669268 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.678983 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr"] Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.761294 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9dd491a2-a155-401a-86fb-04e5fe23984e-apiservice-cert\") pod \"metallb-operator-controller-manager-5fc7dbb978-2m6rr\" (UID: \"9dd491a2-a155-401a-86fb-04e5fe23984e\") " pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.761360 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9dd491a2-a155-401a-86fb-04e5fe23984e-webhook-cert\") pod \"metallb-operator-controller-manager-5fc7dbb978-2m6rr\" (UID: \"9dd491a2-a155-401a-86fb-04e5fe23984e\") " pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.761431 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6ll5\" (UniqueName: \"kubernetes.io/projected/9dd491a2-a155-401a-86fb-04e5fe23984e-kube-api-access-s6ll5\") pod \"metallb-operator-controller-manager-5fc7dbb978-2m6rr\" (UID: \"9dd491a2-a155-401a-86fb-04e5fe23984e\") " pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.862195 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9dd491a2-a155-401a-86fb-04e5fe23984e-apiservice-cert\") pod \"metallb-operator-controller-manager-5fc7dbb978-2m6rr\" (UID: \"9dd491a2-a155-401a-86fb-04e5fe23984e\") " pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.862520 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9dd491a2-a155-401a-86fb-04e5fe23984e-webhook-cert\") pod \"metallb-operator-controller-manager-5fc7dbb978-2m6rr\" (UID: \"9dd491a2-a155-401a-86fb-04e5fe23984e\") " pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.862576 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6ll5\" (UniqueName: \"kubernetes.io/projected/9dd491a2-a155-401a-86fb-04e5fe23984e-kube-api-access-s6ll5\") pod \"metallb-operator-controller-manager-5fc7dbb978-2m6rr\" (UID: \"9dd491a2-a155-401a-86fb-04e5fe23984e\") " pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.867850 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9dd491a2-a155-401a-86fb-04e5fe23984e-webhook-cert\") pod \"metallb-operator-controller-manager-5fc7dbb978-2m6rr\" (UID: \"9dd491a2-a155-401a-86fb-04e5fe23984e\") " pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.868809 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9dd491a2-a155-401a-86fb-04e5fe23984e-apiservice-cert\") pod \"metallb-operator-controller-manager-5fc7dbb978-2m6rr\" (UID: \"9dd491a2-a155-401a-86fb-04e5fe23984e\") " pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.878394 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6ll5\" (UniqueName: \"kubernetes.io/projected/9dd491a2-a155-401a-86fb-04e5fe23984e-kube-api-access-s6ll5\") pod \"metallb-operator-controller-manager-5fc7dbb978-2m6rr\" (UID: \"9dd491a2-a155-401a-86fb-04e5fe23984e\") " pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.988100 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.990702 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk"] Oct 01 16:31:48 crc kubenswrapper[4726]: I1001 16:31:48.992102 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.005509 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk"] Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.065838 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4kzp\" (UniqueName: \"kubernetes.io/projected/f729348c-0bd7-4667-86f9-94e36a9afb11-kube-api-access-h4kzp\") pod \"metallb-operator-webhook-server-d8cc586f4-l28zk\" (UID: \"f729348c-0bd7-4667-86f9-94e36a9afb11\") " pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.066115 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f729348c-0bd7-4667-86f9-94e36a9afb11-webhook-cert\") pod \"metallb-operator-webhook-server-d8cc586f4-l28zk\" (UID: \"f729348c-0bd7-4667-86f9-94e36a9afb11\") " pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.066351 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f729348c-0bd7-4667-86f9-94e36a9afb11-apiservice-cert\") pod \"metallb-operator-webhook-server-d8cc586f4-l28zk\" (UID: \"f729348c-0bd7-4667-86f9-94e36a9afb11\") " pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.167520 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f729348c-0bd7-4667-86f9-94e36a9afb11-apiservice-cert\") pod \"metallb-operator-webhook-server-d8cc586f4-l28zk\" (UID: \"f729348c-0bd7-4667-86f9-94e36a9afb11\") " pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.167947 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4kzp\" (UniqueName: \"kubernetes.io/projected/f729348c-0bd7-4667-86f9-94e36a9afb11-kube-api-access-h4kzp\") pod \"metallb-operator-webhook-server-d8cc586f4-l28zk\" (UID: \"f729348c-0bd7-4667-86f9-94e36a9afb11\") " pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.167978 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f729348c-0bd7-4667-86f9-94e36a9afb11-webhook-cert\") pod \"metallb-operator-webhook-server-d8cc586f4-l28zk\" (UID: \"f729348c-0bd7-4667-86f9-94e36a9afb11\") " pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.173921 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f729348c-0bd7-4667-86f9-94e36a9afb11-webhook-cert\") pod \"metallb-operator-webhook-server-d8cc586f4-l28zk\" (UID: \"f729348c-0bd7-4667-86f9-94e36a9afb11\") " pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.173964 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f729348c-0bd7-4667-86f9-94e36a9afb11-apiservice-cert\") pod \"metallb-operator-webhook-server-d8cc586f4-l28zk\" (UID: \"f729348c-0bd7-4667-86f9-94e36a9afb11\") " pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.185779 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4kzp\" (UniqueName: \"kubernetes.io/projected/f729348c-0bd7-4667-86f9-94e36a9afb11-kube-api-access-h4kzp\") pod \"metallb-operator-webhook-server-d8cc586f4-l28zk\" (UID: \"f729348c-0bd7-4667-86f9-94e36a9afb11\") " pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.326321 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.464781 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr"] Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.799138 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" event={"ID":"9dd491a2-a155-401a-86fb-04e5fe23984e","Type":"ContainerStarted","Data":"f6cae7a985247d2a62a0a985a9402bb717e9c5e51785a2e6622ca01c3a5ea0ab"} Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.799209 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" event={"ID":"9dd491a2-a155-401a-86fb-04e5fe23984e","Type":"ContainerStarted","Data":"7af51c7b5d116fdb627ae983c875803ce2f3a79bf4e18e176f2d90b2b6c466f6"} Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.799282 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.802287 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk"] Oct 01 16:31:49 crc kubenswrapper[4726]: W1001 16:31:49.806014 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf729348c_0bd7_4667_86f9_94e36a9afb11.slice/crio-6ae7fc434ab988e4e0b2b31bc2b883158014d793459728be7b527c0ecf0580ab WatchSource:0}: Error finding container 6ae7fc434ab988e4e0b2b31bc2b883158014d793459728be7b527c0ecf0580ab: Status 404 returned error can't find the container with id 6ae7fc434ab988e4e0b2b31bc2b883158014d793459728be7b527c0ecf0580ab Oct 01 16:31:49 crc kubenswrapper[4726]: I1001 16:31:49.840579 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" podStartSLOduration=1.840555199 podStartE2EDuration="1.840555199s" podCreationTimestamp="2025-10-01 16:31:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:31:49.827953664 +0000 UTC m=+1962.729506281" watchObservedRunningTime="2025-10-01 16:31:49.840555199 +0000 UTC m=+1962.742107776" Oct 01 16:31:50 crc kubenswrapper[4726]: I1001 16:31:50.810847 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" event={"ID":"f729348c-0bd7-4667-86f9-94e36a9afb11","Type":"ContainerStarted","Data":"7ba41fa93c69d134c688116ec87c0fd863109da35fb3bdc296b9bb25b39e898c"} Oct 01 16:31:50 crc kubenswrapper[4726]: I1001 16:31:50.811245 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" event={"ID":"f729348c-0bd7-4667-86f9-94e36a9afb11","Type":"ContainerStarted","Data":"6ae7fc434ab988e4e0b2b31bc2b883158014d793459728be7b527c0ecf0580ab"} Oct 01 16:31:50 crc kubenswrapper[4726]: I1001 16:31:50.811272 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" Oct 01 16:31:50 crc kubenswrapper[4726]: I1001 16:31:50.813158 4726 generic.go:334] "Generic (PLEG): container finished" podID="016c122e-b1fa-43a1-a806-93533e24b8f6" containerID="df3c0d982b9e4d802e28760d1df645b84c6d7054906c1aeb777d85584d9c1b0d" exitCode=0 Oct 01 16:31:50 crc kubenswrapper[4726]: I1001 16:31:50.813252 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" event={"ID":"016c122e-b1fa-43a1-a806-93533e24b8f6","Type":"ContainerDied","Data":"df3c0d982b9e4d802e28760d1df645b84c6d7054906c1aeb777d85584d9c1b0d"} Oct 01 16:31:50 crc kubenswrapper[4726]: I1001 16:31:50.836127 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" podStartSLOduration=2.83610332 podStartE2EDuration="2.83610332s" podCreationTimestamp="2025-10-01 16:31:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:31:50.832539927 +0000 UTC m=+1963.734092524" watchObservedRunningTime="2025-10-01 16:31:50.83610332 +0000 UTC m=+1963.737655897" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.300323 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.450720 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/016c122e-b1fa-43a1-a806-93533e24b8f6-inventory\") pod \"016c122e-b1fa-43a1-a806-93533e24b8f6\" (UID: \"016c122e-b1fa-43a1-a806-93533e24b8f6\") " Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.450885 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/016c122e-b1fa-43a1-a806-93533e24b8f6-ssh-key\") pod \"016c122e-b1fa-43a1-a806-93533e24b8f6\" (UID: \"016c122e-b1fa-43a1-a806-93533e24b8f6\") " Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.450994 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9wfb\" (UniqueName: \"kubernetes.io/projected/016c122e-b1fa-43a1-a806-93533e24b8f6-kube-api-access-f9wfb\") pod \"016c122e-b1fa-43a1-a806-93533e24b8f6\" (UID: \"016c122e-b1fa-43a1-a806-93533e24b8f6\") " Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.456170 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/016c122e-b1fa-43a1-a806-93533e24b8f6-kube-api-access-f9wfb" (OuterVolumeSpecName: "kube-api-access-f9wfb") pod "016c122e-b1fa-43a1-a806-93533e24b8f6" (UID: "016c122e-b1fa-43a1-a806-93533e24b8f6"). InnerVolumeSpecName "kube-api-access-f9wfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.483895 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/016c122e-b1fa-43a1-a806-93533e24b8f6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "016c122e-b1fa-43a1-a806-93533e24b8f6" (UID: "016c122e-b1fa-43a1-a806-93533e24b8f6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.484556 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/016c122e-b1fa-43a1-a806-93533e24b8f6-inventory" (OuterVolumeSpecName: "inventory") pod "016c122e-b1fa-43a1-a806-93533e24b8f6" (UID: "016c122e-b1fa-43a1-a806-93533e24b8f6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.553020 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/016c122e-b1fa-43a1-a806-93533e24b8f6-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.553091 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/016c122e-b1fa-43a1-a806-93533e24b8f6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.553118 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9wfb\" (UniqueName: \"kubernetes.io/projected/016c122e-b1fa-43a1-a806-93533e24b8f6-kube-api-access-f9wfb\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.837607 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" event={"ID":"016c122e-b1fa-43a1-a806-93533e24b8f6","Type":"ContainerDied","Data":"659f81010c67cb4391374bbbcfb54ad185b52778f800484251b4130635dfe147"} Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.837649 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="659f81010c67cb4391374bbbcfb54ad185b52778f800484251b4130635dfe147" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.837657 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.955666 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j"] Oct 01 16:31:52 crc kubenswrapper[4726]: E1001 16:31:52.956471 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="016c122e-b1fa-43a1-a806-93533e24b8f6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.956497 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="016c122e-b1fa-43a1-a806-93533e24b8f6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.956758 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="016c122e-b1fa-43a1-a806-93533e24b8f6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.957555 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.960209 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.961324 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.961548 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.961743 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.961792 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.961833 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2d6v\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-kube-api-access-k2d6v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.961871 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.961911 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.961953 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.962090 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.962124 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.962223 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.962264 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.962303 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.962337 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.962389 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.962420 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.963560 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.963746 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.963894 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.964759 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j"] Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.974115 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 01 16:31:52 crc kubenswrapper[4726]: I1001 16:31:52.976870 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.064036 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.064185 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.064232 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.064255 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.064321 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.064363 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.064399 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.064429 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.064474 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.064508 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.064533 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.064563 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.064639 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2d6v\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-kube-api-access-k2d6v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.064676 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.069026 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.069076 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.071012 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.072322 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.072429 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.072532 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.073118 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.073486 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.074305 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.074335 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.076575 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.079158 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.079863 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.084027 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2d6v\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-kube-api-access-k2d6v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.289150 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:31:53 crc kubenswrapper[4726]: I1001 16:31:53.911181 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j"] Oct 01 16:31:54 crc kubenswrapper[4726]: I1001 16:31:54.857350 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" event={"ID":"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2","Type":"ContainerStarted","Data":"dfbdfbb8e3069c25965a38d5fe75c4599f7ee394c95eeee223647dae2a8b7797"} Oct 01 16:31:54 crc kubenswrapper[4726]: I1001 16:31:54.857651 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" event={"ID":"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2","Type":"ContainerStarted","Data":"0897e1977b685b34baccb25a779ea9bd84c45f181abed3c372d3b9aafd9fe38d"} Oct 01 16:31:54 crc kubenswrapper[4726]: I1001 16:31:54.888968 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" podStartSLOduration=2.451901659 podStartE2EDuration="2.888942419s" podCreationTimestamp="2025-10-01 16:31:52 +0000 UTC" firstStartedPulling="2025-10-01 16:31:53.891080882 +0000 UTC m=+1966.792633459" lastFinishedPulling="2025-10-01 16:31:54.328121622 +0000 UTC m=+1967.229674219" observedRunningTime="2025-10-01 16:31:54.881187494 +0000 UTC m=+1967.782740091" watchObservedRunningTime="2025-10-01 16:31:54.888942419 +0000 UTC m=+1967.790494996" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.454880 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/frr-k8s-r4djs"] Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.455782 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-r4djs" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="controller" containerID="cri-o://a1374a43ccb81a9a3f97068c3ca67cf90019219e90d280cff217e2b3c54782ef" gracePeriod=2 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.456166 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-r4djs" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="kube-rbac-proxy" containerID="cri-o://52c0e14d3038a74f8536411a7900dff96b2a9ae673439bf230ae87996a239754" gracePeriod=2 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.456217 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-r4djs" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="reloader" containerID="cri-o://33e1649f8c3dabd843a405caf4386c07f38d0c68d7380a48b0cbc7e7e679c3cc" gracePeriod=2 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.456326 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-r4djs" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="frr-metrics" containerID="cri-o://abd2a7e13f99d37144d0076c20361976ce336847645daf450d24449bbadfe0ab" gracePeriod=2 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.456371 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-r4djs" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="kube-rbac-proxy-frr" containerID="cri-o://21e25a53c4de4535261a28d2622d716c55aa1ae4b3e304ee4cf0c9689471cb6c" gracePeriod=2 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.456449 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-r4djs" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="frr" containerID="cri-o://b15081c4adf77afe809fbd87b422a48727f4320cbfa77d910c72a963c0b485f6" gracePeriod=2 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.484390 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/frr-k8s-r4djs"] Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.498927 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2"] Oct 01 16:31:58 crc kubenswrapper[4726]: E1001 16:31:58.499488 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="cp-reloader" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.499511 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="cp-reloader" Oct 01 16:31:58 crc kubenswrapper[4726]: E1001 16:31:58.499531 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="controller" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.499541 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="controller" Oct 01 16:31:58 crc kubenswrapper[4726]: E1001 16:31:58.499560 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="kube-rbac-proxy" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.499569 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="kube-rbac-proxy" Oct 01 16:31:58 crc kubenswrapper[4726]: E1001 16:31:58.499591 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="kube-rbac-proxy-frr" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.499600 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="kube-rbac-proxy-frr" Oct 01 16:31:58 crc kubenswrapper[4726]: E1001 16:31:58.499613 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="frr-metrics" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.499621 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="frr-metrics" Oct 01 16:31:58 crc kubenswrapper[4726]: E1001 16:31:58.499661 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="reloader" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.499669 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="reloader" Oct 01 16:31:58 crc kubenswrapper[4726]: E1001 16:31:58.499681 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="frr" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.499690 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="frr" Oct 01 16:31:58 crc kubenswrapper[4726]: E1001 16:31:58.499709 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="cp-frr-files" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.499719 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="cp-frr-files" Oct 01 16:31:58 crc kubenswrapper[4726]: E1001 16:31:58.499731 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="cp-metrics" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.499738 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="cp-metrics" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.499995 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="frr-metrics" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.500016 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="kube-rbac-proxy" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.500037 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="reloader" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.500070 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="controller" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.500092 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="kube-rbac-proxy-frr" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.500110 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerName="frr" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.501067 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.536979 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2"] Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.561131 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-xqs68"] Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.568955 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.599937 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkz4m\" (UniqueName: \"kubernetes.io/projected/678af4b7-7942-40da-8273-de583fd22666-kube-api-access-tkz4m\") pod \"frr-k8s-webhook-server-64bf5d555-mqdn2\" (UID: \"678af4b7-7942-40da-8273-de583fd22666\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.600008 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/678af4b7-7942-40da-8273-de583fd22666-cert\") pod \"frr-k8s-webhook-server-64bf5d555-mqdn2\" (UID: \"678af4b7-7942-40da-8273-de583fd22666\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.701490 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-reloader\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.701541 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-frr-conf\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.701599 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-frr-sockets\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.701625 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-frr-startup\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.701667 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7mkn\" (UniqueName: \"kubernetes.io/projected/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-kube-api-access-z7mkn\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.701707 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkz4m\" (UniqueName: \"kubernetes.io/projected/678af4b7-7942-40da-8273-de583fd22666-kube-api-access-tkz4m\") pod \"frr-k8s-webhook-server-64bf5d555-mqdn2\" (UID: \"678af4b7-7942-40da-8273-de583fd22666\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.701723 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-metrics\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.701755 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/678af4b7-7942-40da-8273-de583fd22666-cert\") pod \"frr-k8s-webhook-server-64bf5d555-mqdn2\" (UID: \"678af4b7-7942-40da-8273-de583fd22666\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.701785 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-metrics-certs\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.706134 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/speaker-9frd4"] Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.706390 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/speaker-9frd4" podUID="08f6790e-97a0-48c4-815d-cb6d57fa935c" containerName="speaker" containerID="cri-o://e7934c6514acc3b6103da5da3fe1f2fcb7d1b84b3ce5f8e705b31de97f4d0620" gracePeriod=2 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.706758 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/speaker-9frd4" podUID="08f6790e-97a0-48c4-815d-cb6d57fa935c" containerName="kube-rbac-proxy" containerID="cri-o://f0050ea6732902ce8c291ef445507eee0396924e50670b6996c61d4eaa7dc647" gracePeriod=2 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.731907 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/678af4b7-7942-40da-8273-de583fd22666-cert\") pod \"frr-k8s-webhook-server-64bf5d555-mqdn2\" (UID: \"678af4b7-7942-40da-8273-de583fd22666\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.750561 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/speaker-9frd4"] Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.752744 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkz4m\" (UniqueName: \"kubernetes.io/projected/678af4b7-7942-40da-8273-de583fd22666-kube-api-access-tkz4m\") pod \"frr-k8s-webhook-server-64bf5d555-mqdn2\" (UID: \"678af4b7-7942-40da-8273-de583fd22666\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.771930 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-8tdc4"] Oct 01 16:31:58 crc kubenswrapper[4726]: E1001 16:31:58.773159 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08f6790e-97a0-48c4-815d-cb6d57fa935c" containerName="speaker" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.773176 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="08f6790e-97a0-48c4-815d-cb6d57fa935c" containerName="speaker" Oct 01 16:31:58 crc kubenswrapper[4726]: E1001 16:31:58.773198 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08f6790e-97a0-48c4-815d-cb6d57fa935c" containerName="kube-rbac-proxy" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.773205 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="08f6790e-97a0-48c4-815d-cb6d57fa935c" containerName="kube-rbac-proxy" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.773436 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="08f6790e-97a0-48c4-815d-cb6d57fa935c" containerName="speaker" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.773448 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="08f6790e-97a0-48c4-815d-cb6d57fa935c" containerName="kube-rbac-proxy" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.774413 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-8tdc4" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.807198 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-metrics-certs\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.807539 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-reloader\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.807577 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-frr-conf\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.807633 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-frr-sockets\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.807661 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-frr-startup\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.807707 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7mkn\" (UniqueName: \"kubernetes.io/projected/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-kube-api-access-z7mkn\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.807743 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-metrics\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.808168 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-metrics\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.808372 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-frr-conf\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.809259 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-frr-startup\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.827498 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-frr-sockets\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.837459 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-metrics-certs\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.839914 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-reloader\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.847623 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-kk7mh"] Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.850863 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-kk7mh" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.856543 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.877710 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7mkn\" (UniqueName: \"kubernetes.io/projected/9b9a95f4-3ea1-45a0-8884-b6dc26ad5895-kube-api-access-z7mkn\") pod \"frr-k8s-xqs68\" (UID: \"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895\") " pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.902941 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-kk7mh"] Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.909903 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/087620f7-11b5-49b6-a8b7-30a75a5196cb-memberlist\") pod \"speaker-8tdc4\" (UID: \"087620f7-11b5-49b6-a8b7-30a75a5196cb\") " pod="metallb-system/speaker-8tdc4" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.909983 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/087620f7-11b5-49b6-a8b7-30a75a5196cb-metrics-certs\") pod \"speaker-8tdc4\" (UID: \"087620f7-11b5-49b6-a8b7-30a75a5196cb\") " pod="metallb-system/speaker-8tdc4" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.910127 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/087620f7-11b5-49b6-a8b7-30a75a5196cb-metallb-excludel2\") pod \"speaker-8tdc4\" (UID: \"087620f7-11b5-49b6-a8b7-30a75a5196cb\") " pod="metallb-system/speaker-8tdc4" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.910166 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84c4c\" (UniqueName: \"kubernetes.io/projected/087620f7-11b5-49b6-a8b7-30a75a5196cb-kube-api-access-84c4c\") pod \"speaker-8tdc4\" (UID: \"087620f7-11b5-49b6-a8b7-30a75a5196cb\") " pod="metallb-system/speaker-8tdc4" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.932678 4726 generic.go:334] "Generic (PLEG): container finished" podID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerID="21e25a53c4de4535261a28d2622d716c55aa1ae4b3e304ee4cf0c9689471cb6c" exitCode=0 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.932707 4726 generic.go:334] "Generic (PLEG): container finished" podID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerID="52c0e14d3038a74f8536411a7900dff96b2a9ae673439bf230ae87996a239754" exitCode=0 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.932716 4726 generic.go:334] "Generic (PLEG): container finished" podID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerID="abd2a7e13f99d37144d0076c20361976ce336847645daf450d24449bbadfe0ab" exitCode=143 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.932723 4726 generic.go:334] "Generic (PLEG): container finished" podID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerID="33e1649f8c3dabd843a405caf4386c07f38d0c68d7380a48b0cbc7e7e679c3cc" exitCode=0 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.932731 4726 generic.go:334] "Generic (PLEG): container finished" podID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerID="b15081c4adf77afe809fbd87b422a48727f4320cbfa77d910c72a963c0b485f6" exitCode=143 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.932738 4726 generic.go:334] "Generic (PLEG): container finished" podID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" containerID="a1374a43ccb81a9a3f97068c3ca67cf90019219e90d280cff217e2b3c54782ef" exitCode=0 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.932794 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c19a6336e132b681c509e5f0efcbfb8968d35617274378deb4b650eb438efc7" Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.936931 4726 generic.go:334] "Generic (PLEG): container finished" podID="08f6790e-97a0-48c4-815d-cb6d57fa935c" containerID="f0050ea6732902ce8c291ef445507eee0396924e50670b6996c61d4eaa7dc647" exitCode=0 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.936966 4726 generic.go:334] "Generic (PLEG): container finished" podID="08f6790e-97a0-48c4-815d-cb6d57fa935c" containerID="e7934c6514acc3b6103da5da3fe1f2fcb7d1b84b3ce5f8e705b31de97f4d0620" exitCode=0 Oct 01 16:31:58 crc kubenswrapper[4726]: I1001 16:31:58.978353 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-r4djs" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.006667 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-xqs68" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.017264 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnwwt\" (UniqueName: \"kubernetes.io/projected/1c256ab6-be12-4843-a072-f6bb861c9740-kube-api-access-gnwwt\") pod \"controller-68d546b9d8-kk7mh\" (UID: \"1c256ab6-be12-4843-a072-f6bb861c9740\") " pod="metallb-system/controller-68d546b9d8-kk7mh" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.017332 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/087620f7-11b5-49b6-a8b7-30a75a5196cb-memberlist\") pod \"speaker-8tdc4\" (UID: \"087620f7-11b5-49b6-a8b7-30a75a5196cb\") " pod="metallb-system/speaker-8tdc4" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.017373 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c256ab6-be12-4843-a072-f6bb861c9740-cert\") pod \"controller-68d546b9d8-kk7mh\" (UID: \"1c256ab6-be12-4843-a072-f6bb861c9740\") " pod="metallb-system/controller-68d546b9d8-kk7mh" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.017421 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1c256ab6-be12-4843-a072-f6bb861c9740-metrics-certs\") pod \"controller-68d546b9d8-kk7mh\" (UID: \"1c256ab6-be12-4843-a072-f6bb861c9740\") " pod="metallb-system/controller-68d546b9d8-kk7mh" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.017448 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/087620f7-11b5-49b6-a8b7-30a75a5196cb-metrics-certs\") pod \"speaker-8tdc4\" (UID: \"087620f7-11b5-49b6-a8b7-30a75a5196cb\") " pod="metallb-system/speaker-8tdc4" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.017508 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/087620f7-11b5-49b6-a8b7-30a75a5196cb-metallb-excludel2\") pod \"speaker-8tdc4\" (UID: \"087620f7-11b5-49b6-a8b7-30a75a5196cb\") " pod="metallb-system/speaker-8tdc4" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.017557 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84c4c\" (UniqueName: \"kubernetes.io/projected/087620f7-11b5-49b6-a8b7-30a75a5196cb-kube-api-access-84c4c\") pod \"speaker-8tdc4\" (UID: \"087620f7-11b5-49b6-a8b7-30a75a5196cb\") " pod="metallb-system/speaker-8tdc4" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.025230 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/087620f7-11b5-49b6-a8b7-30a75a5196cb-metallb-excludel2\") pod \"speaker-8tdc4\" (UID: \"087620f7-11b5-49b6-a8b7-30a75a5196cb\") " pod="metallb-system/speaker-8tdc4" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.037435 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/087620f7-11b5-49b6-a8b7-30a75a5196cb-metrics-certs\") pod \"speaker-8tdc4\" (UID: \"087620f7-11b5-49b6-a8b7-30a75a5196cb\") " pod="metallb-system/speaker-8tdc4" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.038219 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/087620f7-11b5-49b6-a8b7-30a75a5196cb-memberlist\") pod \"speaker-8tdc4\" (UID: \"087620f7-11b5-49b6-a8b7-30a75a5196cb\") " pod="metallb-system/speaker-8tdc4" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.039837 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9frd4" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.044954 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84c4c\" (UniqueName: \"kubernetes.io/projected/087620f7-11b5-49b6-a8b7-30a75a5196cb-kube-api-access-84c4c\") pod \"speaker-8tdc4\" (UID: \"087620f7-11b5-49b6-a8b7-30a75a5196cb\") " pod="metallb-system/speaker-8tdc4" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.118371 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-metrics-certs\") pod \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.118467 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-metrics\") pod \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.118723 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-reloader\") pod \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.118756 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcgnw\" (UniqueName: \"kubernetes.io/projected/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-kube-api-access-lcgnw\") pod \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.118802 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-startup\") pod \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.118845 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-conf\") pod \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.118955 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-sockets\") pod \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\" (UID: \"f08bbaf6-372d-43ab-aa18-d6870e95fe4c\") " Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.119362 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1c256ab6-be12-4843-a072-f6bb861c9740-metrics-certs\") pod \"controller-68d546b9d8-kk7mh\" (UID: \"1c256ab6-be12-4843-a072-f6bb861c9740\") " pod="metallb-system/controller-68d546b9d8-kk7mh" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.119622 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnwwt\" (UniqueName: \"kubernetes.io/projected/1c256ab6-be12-4843-a072-f6bb861c9740-kube-api-access-gnwwt\") pod \"controller-68d546b9d8-kk7mh\" (UID: \"1c256ab6-be12-4843-a072-f6bb861c9740\") " pod="metallb-system/controller-68d546b9d8-kk7mh" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.119678 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c256ab6-be12-4843-a072-f6bb861c9740-cert\") pod \"controller-68d546b9d8-kk7mh\" (UID: \"1c256ab6-be12-4843-a072-f6bb861c9740\") " pod="metallb-system/controller-68d546b9d8-kk7mh" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.120359 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-startup" (OuterVolumeSpecName: "frr-startup") pod "f08bbaf6-372d-43ab-aa18-d6870e95fe4c" (UID: "f08bbaf6-372d-43ab-aa18-d6870e95fe4c"). InnerVolumeSpecName "frr-startup". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.121457 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-sockets" (OuterVolumeSpecName: "frr-sockets") pod "f08bbaf6-372d-43ab-aa18-d6870e95fe4c" (UID: "f08bbaf6-372d-43ab-aa18-d6870e95fe4c"). InnerVolumeSpecName "frr-sockets". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.123177 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "f08bbaf6-372d-43ab-aa18-d6870e95fe4c" (UID: "f08bbaf6-372d-43ab-aa18-d6870e95fe4c"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.125039 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-reloader" (OuterVolumeSpecName: "reloader") pod "f08bbaf6-372d-43ab-aa18-d6870e95fe4c" (UID: "f08bbaf6-372d-43ab-aa18-d6870e95fe4c"). InnerVolumeSpecName "reloader". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.126606 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-kube-api-access-lcgnw" (OuterVolumeSpecName: "kube-api-access-lcgnw") pod "f08bbaf6-372d-43ab-aa18-d6870e95fe4c" (UID: "f08bbaf6-372d-43ab-aa18-d6870e95fe4c"). InnerVolumeSpecName "kube-api-access-lcgnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.131546 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c256ab6-be12-4843-a072-f6bb861c9740-cert\") pod \"controller-68d546b9d8-kk7mh\" (UID: \"1c256ab6-be12-4843-a072-f6bb861c9740\") " pod="metallb-system/controller-68d546b9d8-kk7mh" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.131792 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-conf" (OuterVolumeSpecName: "frr-conf") pod "f08bbaf6-372d-43ab-aa18-d6870e95fe4c" (UID: "f08bbaf6-372d-43ab-aa18-d6870e95fe4c"). InnerVolumeSpecName "frr-conf". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.135724 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1c256ab6-be12-4843-a072-f6bb861c9740-metrics-certs\") pod \"controller-68d546b9d8-kk7mh\" (UID: \"1c256ab6-be12-4843-a072-f6bb861c9740\") " pod="metallb-system/controller-68d546b9d8-kk7mh" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.136639 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-metrics" (OuterVolumeSpecName: "metrics") pod "f08bbaf6-372d-43ab-aa18-d6870e95fe4c" (UID: "f08bbaf6-372d-43ab-aa18-d6870e95fe4c"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.148475 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnwwt\" (UniqueName: \"kubernetes.io/projected/1c256ab6-be12-4843-a072-f6bb861c9740-kube-api-access-gnwwt\") pod \"controller-68d546b9d8-kk7mh\" (UID: \"1c256ab6-be12-4843-a072-f6bb861c9740\") " pod="metallb-system/controller-68d546b9d8-kk7mh" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.221827 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/08f6790e-97a0-48c4-815d-cb6d57fa935c-metallb-excludel2\") pod \"08f6790e-97a0-48c4-815d-cb6d57fa935c\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.221932 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g77jf\" (UniqueName: \"kubernetes.io/projected/08f6790e-97a0-48c4-815d-cb6d57fa935c-kube-api-access-g77jf\") pod \"08f6790e-97a0-48c4-815d-cb6d57fa935c\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.222000 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-metrics-certs\") pod \"08f6790e-97a0-48c4-815d-cb6d57fa935c\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.222326 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-memberlist\") pod \"08f6790e-97a0-48c4-815d-cb6d57fa935c\" (UID: \"08f6790e-97a0-48c4-815d-cb6d57fa935c\") " Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.222934 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08f6790e-97a0-48c4-815d-cb6d57fa935c-metallb-excludel2" (OuterVolumeSpecName: "metallb-excludel2") pod "08f6790e-97a0-48c4-815d-cb6d57fa935c" (UID: "08f6790e-97a0-48c4-815d-cb6d57fa935c"). InnerVolumeSpecName "metallb-excludel2". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.223213 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.223244 4726 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-metrics\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.223259 4726 reconciler_common.go:293] "Volume detached for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-reloader\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.223292 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcgnw\" (UniqueName: \"kubernetes.io/projected/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-kube-api-access-lcgnw\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.223303 4726 reconciler_common.go:293] "Volume detached for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-startup\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.223313 4726 reconciler_common.go:293] "Volume detached for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-conf\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.223324 4726 reconciler_common.go:293] "Volume detached for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/08f6790e-97a0-48c4-815d-cb6d57fa935c-metallb-excludel2\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.223382 4726 reconciler_common.go:293] "Volume detached for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f08bbaf6-372d-43ab-aa18-d6870e95fe4c-frr-sockets\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.226376 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08f6790e-97a0-48c4-815d-cb6d57fa935c-kube-api-access-g77jf" (OuterVolumeSpecName: "kube-api-access-g77jf") pod "08f6790e-97a0-48c4-815d-cb6d57fa935c" (UID: "08f6790e-97a0-48c4-815d-cb6d57fa935c"). InnerVolumeSpecName "kube-api-access-g77jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.226800 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "08f6790e-97a0-48c4-815d-cb6d57fa935c" (UID: "08f6790e-97a0-48c4-815d-cb6d57fa935c"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.226859 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-memberlist" (OuterVolumeSpecName: "memberlist") pod "08f6790e-97a0-48c4-815d-cb6d57fa935c" (UID: "08f6790e-97a0-48c4-815d-cb6d57fa935c"). InnerVolumeSpecName "memberlist". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.250810 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-8tdc4" Oct 01 16:31:59 crc kubenswrapper[4726]: W1001 16:31:59.283557 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod087620f7_11b5_49b6_a8b7_30a75a5196cb.slice/crio-a2ec486cbfbf548c7946c0f2349297499754a9cfaeed8395257716257b75c975 WatchSource:0}: Error finding container a2ec486cbfbf548c7946c0f2349297499754a9cfaeed8395257716257b75c975: Status 404 returned error can't find the container with id a2ec486cbfbf548c7946c0f2349297499754a9cfaeed8395257716257b75c975 Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.322209 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-kk7mh" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.325744 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g77jf\" (UniqueName: \"kubernetes.io/projected/08f6790e-97a0-48c4-815d-cb6d57fa935c-kube-api-access-g77jf\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.325782 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.325792 4726 reconciler_common.go:293] "Volume detached for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/08f6790e-97a0-48c4-815d-cb6d57fa935c-memberlist\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:59 crc kubenswrapper[4726]: W1001 16:31:59.413935 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod678af4b7_7942_40da_8273_de583fd22666.slice/crio-0f18bbbdc2ac21c79db1988d1fbcefa80b80e8d9764a3734ccf893f29cefe4b1 WatchSource:0}: Error finding container 0f18bbbdc2ac21c79db1988d1fbcefa80b80e8d9764a3734ccf893f29cefe4b1: Status 404 returned error can't find the container with id 0f18bbbdc2ac21c79db1988d1fbcefa80b80e8d9764a3734ccf893f29cefe4b1 Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.417920 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2"] Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.838526 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08f6790e-97a0-48c4-815d-cb6d57fa935c" path="/var/lib/kubelet/pods/08f6790e-97a0-48c4-815d-cb6d57fa935c/volumes" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.842632 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f08bbaf6-372d-43ab-aa18-d6870e95fe4c" path="/var/lib/kubelet/pods/f08bbaf6-372d-43ab-aa18-d6870e95fe4c/volumes" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.845766 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-kk7mh"] Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.960222 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xqs68" event={"ID":"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895","Type":"ContainerStarted","Data":"79f99fe7eb2e19b6c796b4733e174b4eac96a97945a47eb5aa93deefdf2fe067"} Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.981666 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-kk7mh" event={"ID":"1c256ab6-be12-4843-a072-f6bb861c9740","Type":"ContainerStarted","Data":"74328bc31669c5394fe87293667aaf23b2aee89ae8751f0c4bc1dc2584960ff3"} Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.983480 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2" event={"ID":"678af4b7-7942-40da-8273-de583fd22666","Type":"ContainerStarted","Data":"0f18bbbdc2ac21c79db1988d1fbcefa80b80e8d9764a3734ccf893f29cefe4b1"} Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.984806 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8tdc4" event={"ID":"087620f7-11b5-49b6-a8b7-30a75a5196cb","Type":"ContainerStarted","Data":"0f3f8c8bf1db6d1d3245f82b57b0007c95e803b06a04721870baa490ce30f4cc"} Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.984842 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8tdc4" event={"ID":"087620f7-11b5-49b6-a8b7-30a75a5196cb","Type":"ContainerStarted","Data":"a2ec486cbfbf548c7946c0f2349297499754a9cfaeed8395257716257b75c975"} Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.986504 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-r4djs" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.987489 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9frd4" Oct 01 16:31:59 crc kubenswrapper[4726]: I1001 16:31:59.987546 4726 scope.go:117] "RemoveContainer" containerID="f0050ea6732902ce8c291ef445507eee0396924e50670b6996c61d4eaa7dc647" Oct 01 16:32:00 crc kubenswrapper[4726]: I1001 16:32:00.028036 4726 scope.go:117] "RemoveContainer" containerID="e7934c6514acc3b6103da5da3fe1f2fcb7d1b84b3ce5f8e705b31de97f4d0620" Oct 01 16:32:01 crc kubenswrapper[4726]: I1001 16:32:01.000835 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8tdc4" event={"ID":"087620f7-11b5-49b6-a8b7-30a75a5196cb","Type":"ContainerStarted","Data":"959116e20d7b9c12c03ac62ddf86929a842f320a71bbe83271a86f58a40c5715"} Oct 01 16:32:01 crc kubenswrapper[4726]: I1001 16:32:01.001208 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-8tdc4" Oct 01 16:32:01 crc kubenswrapper[4726]: I1001 16:32:01.006680 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-kk7mh" event={"ID":"1c256ab6-be12-4843-a072-f6bb861c9740","Type":"ContainerStarted","Data":"f449c1614129717dd79072e8aa97533a93bf72a2972172268cb931a7610b8afb"} Oct 01 16:32:01 crc kubenswrapper[4726]: I1001 16:32:01.006712 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-kk7mh" event={"ID":"1c256ab6-be12-4843-a072-f6bb861c9740","Type":"ContainerStarted","Data":"1edf79da77ceab29b3e247988f0019d08cd9954df5619782d323b6e0298af3f5"} Oct 01 16:32:01 crc kubenswrapper[4726]: I1001 16:32:01.007721 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-kk7mh" Oct 01 16:32:01 crc kubenswrapper[4726]: I1001 16:32:01.027129 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-8tdc4" podStartSLOduration=3.027104723 podStartE2EDuration="3.027104723s" podCreationTimestamp="2025-10-01 16:31:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:32:01.018527404 +0000 UTC m=+1973.920080001" watchObservedRunningTime="2025-10-01 16:32:01.027104723 +0000 UTC m=+1973.928657300" Oct 01 16:32:01 crc kubenswrapper[4726]: I1001 16:32:01.051374 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-kk7mh" podStartSLOduration=3.051320505 podStartE2EDuration="3.051320505s" podCreationTimestamp="2025-10-01 16:31:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:32:01.046960458 +0000 UTC m=+1973.948513045" watchObservedRunningTime="2025-10-01 16:32:01.051320505 +0000 UTC m=+1973.952873082" Oct 01 16:32:06 crc kubenswrapper[4726]: I1001 16:32:06.074621 4726 generic.go:334] "Generic (PLEG): container finished" podID="9b9a95f4-3ea1-45a0-8884-b6dc26ad5895" containerID="810c836563f37e0e256eefa24e9c998fad34cb189aa89646d41e652caf673541" exitCode=0 Oct 01 16:32:06 crc kubenswrapper[4726]: I1001 16:32:06.074728 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xqs68" event={"ID":"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895","Type":"ContainerDied","Data":"810c836563f37e0e256eefa24e9c998fad34cb189aa89646d41e652caf673541"} Oct 01 16:32:06 crc kubenswrapper[4726]: I1001 16:32:06.077297 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2" event={"ID":"678af4b7-7942-40da-8273-de583fd22666","Type":"ContainerStarted","Data":"22e5fe02fc1a2a298300966400cd20afe378193685515209189efb09e679a321"} Oct 01 16:32:06 crc kubenswrapper[4726]: I1001 16:32:06.077455 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2" Oct 01 16:32:06 crc kubenswrapper[4726]: I1001 16:32:06.126360 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2" podStartSLOduration=1.979547024 podStartE2EDuration="8.126342237s" podCreationTimestamp="2025-10-01 16:31:58 +0000 UTC" firstStartedPulling="2025-10-01 16:31:59.417076018 +0000 UTC m=+1972.318628595" lastFinishedPulling="2025-10-01 16:32:05.563871231 +0000 UTC m=+1978.465423808" observedRunningTime="2025-10-01 16:32:06.118394797 +0000 UTC m=+1979.019947384" watchObservedRunningTime="2025-10-01 16:32:06.126342237 +0000 UTC m=+1979.027894814" Oct 01 16:32:07 crc kubenswrapper[4726]: I1001 16:32:07.089377 4726 generic.go:334] "Generic (PLEG): container finished" podID="9b9a95f4-3ea1-45a0-8884-b6dc26ad5895" containerID="7b93da820e4898bc81da80b11fdf067fc4e6bff9e0d4013d9a41d8d8fc9da164" exitCode=0 Oct 01 16:32:07 crc kubenswrapper[4726]: I1001 16:32:07.089453 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xqs68" event={"ID":"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895","Type":"ContainerDied","Data":"7b93da820e4898bc81da80b11fdf067fc4e6bff9e0d4013d9a41d8d8fc9da164"} Oct 01 16:32:08 crc kubenswrapper[4726]: I1001 16:32:08.101415 4726 generic.go:334] "Generic (PLEG): container finished" podID="9b9a95f4-3ea1-45a0-8884-b6dc26ad5895" containerID="1a39a75b7030679b6774497b8bf9a3496c5341538209797fd784ecd88b34c7ae" exitCode=0 Oct 01 16:32:08 crc kubenswrapper[4726]: I1001 16:32:08.101632 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xqs68" event={"ID":"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895","Type":"ContainerDied","Data":"1a39a75b7030679b6774497b8bf9a3496c5341538209797fd784ecd88b34c7ae"} Oct 01 16:32:09 crc kubenswrapper[4726]: I1001 16:32:09.161246 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xqs68" event={"ID":"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895","Type":"ContainerStarted","Data":"517ffbf8ecf57328fd95a2b3dfec99b7f7ae3950a79a247ba8ffd8ce46d3d2b5"} Oct 01 16:32:09 crc kubenswrapper[4726]: I1001 16:32:09.162372 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xqs68" event={"ID":"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895","Type":"ContainerStarted","Data":"b76d3742db567248dbdb5f3e4c637112c068a4b4e65ae33a88d9f25e9b40f261"} Oct 01 16:32:09 crc kubenswrapper[4726]: I1001 16:32:09.162393 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xqs68" event={"ID":"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895","Type":"ContainerStarted","Data":"86206a543051cbf04a0366ed78040a8369d516b4c1bc5f1056ef68978d52e89d"} Oct 01 16:32:09 crc kubenswrapper[4726]: I1001 16:32:09.162404 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xqs68" event={"ID":"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895","Type":"ContainerStarted","Data":"072d5ecb00198bffaac93a8eee83da01f008eee1649ec15aac962e2895419581"} Oct 01 16:32:09 crc kubenswrapper[4726]: I1001 16:32:09.255661 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-8tdc4" Oct 01 16:32:09 crc kubenswrapper[4726]: I1001 16:32:09.330222 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-d8cc586f4-l28zk" Oct 01 16:32:09 crc kubenswrapper[4726]: I1001 16:32:09.390042 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l"] Oct 01 16:32:09 crc kubenswrapper[4726]: I1001 16:32:09.390306 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" podUID="46a1d32f-b37b-48cb-880f-6cbb959a7b3a" containerName="webhook-server" containerID="cri-o://6b34947dbaee53f404e0aa6070a1f595e64dc9e43284e019093d0a9634b1048f" gracePeriod=2 Oct 01 16:32:09 crc kubenswrapper[4726]: I1001 16:32:09.408534 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l"] Oct 01 16:32:09 crc kubenswrapper[4726]: I1001 16:32:09.892910 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.058067 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-apiservice-cert\") pod \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\" (UID: \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\") " Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.058177 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvv8b\" (UniqueName: \"kubernetes.io/projected/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-kube-api-access-qvv8b\") pod \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\" (UID: \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\") " Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.058445 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-webhook-cert\") pod \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\" (UID: \"46a1d32f-b37b-48cb-880f-6cbb959a7b3a\") " Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.063873 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "46a1d32f-b37b-48cb-880f-6cbb959a7b3a" (UID: "46a1d32f-b37b-48cb-880f-6cbb959a7b3a"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.064202 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-kube-api-access-qvv8b" (OuterVolumeSpecName: "kube-api-access-qvv8b") pod "46a1d32f-b37b-48cb-880f-6cbb959a7b3a" (UID: "46a1d32f-b37b-48cb-880f-6cbb959a7b3a"). InnerVolumeSpecName "kube-api-access-qvv8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.066287 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "46a1d32f-b37b-48cb-880f-6cbb959a7b3a" (UID: "46a1d32f-b37b-48cb-880f-6cbb959a7b3a"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.161308 4726 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.161356 4726 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.161370 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvv8b\" (UniqueName: \"kubernetes.io/projected/46a1d32f-b37b-48cb-880f-6cbb959a7b3a-kube-api-access-qvv8b\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.177996 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xqs68" event={"ID":"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895","Type":"ContainerStarted","Data":"cbfbd3f6112b3986ff12e999bc044543f9f0c33e1bf497e2b61879eb9309a79c"} Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.178040 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xqs68" event={"ID":"9b9a95f4-3ea1-45a0-8884-b6dc26ad5895","Type":"ContainerStarted","Data":"bbcb82a56e6bfc30a8bc6fb934d8aa07f8609350cee09187dbd31790cbe63495"} Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.178513 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-xqs68" Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.181971 4726 generic.go:334] "Generic (PLEG): container finished" podID="46a1d32f-b37b-48cb-880f-6cbb959a7b3a" containerID="6b34947dbaee53f404e0aa6070a1f595e64dc9e43284e019093d0a9634b1048f" exitCode=0 Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.182021 4726 scope.go:117] "RemoveContainer" containerID="6b34947dbaee53f404e0aa6070a1f595e64dc9e43284e019093d0a9634b1048f" Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.182190 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-55b5c94945-7tz8l" Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.209377 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-xqs68" podStartSLOduration=5.846803655 podStartE2EDuration="12.209358753s" podCreationTimestamp="2025-10-01 16:31:58 +0000 UTC" firstStartedPulling="2025-10-01 16:31:59.173774415 +0000 UTC m=+1972.075326992" lastFinishedPulling="2025-10-01 16:32:05.536329513 +0000 UTC m=+1978.437882090" observedRunningTime="2025-10-01 16:32:10.201899007 +0000 UTC m=+1983.103451594" watchObservedRunningTime="2025-10-01 16:32:10.209358753 +0000 UTC m=+1983.110911330" Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.213751 4726 scope.go:117] "RemoveContainer" containerID="6b34947dbaee53f404e0aa6070a1f595e64dc9e43284e019093d0a9634b1048f" Oct 01 16:32:10 crc kubenswrapper[4726]: E1001 16:32:10.214357 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b34947dbaee53f404e0aa6070a1f595e64dc9e43284e019093d0a9634b1048f\": container with ID starting with 6b34947dbaee53f404e0aa6070a1f595e64dc9e43284e019093d0a9634b1048f not found: ID does not exist" containerID="6b34947dbaee53f404e0aa6070a1f595e64dc9e43284e019093d0a9634b1048f" Oct 01 16:32:10 crc kubenswrapper[4726]: I1001 16:32:10.214401 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b34947dbaee53f404e0aa6070a1f595e64dc9e43284e019093d0a9634b1048f"} err="failed to get container status \"6b34947dbaee53f404e0aa6070a1f595e64dc9e43284e019093d0a9634b1048f\": rpc error: code = NotFound desc = could not find container \"6b34947dbaee53f404e0aa6070a1f595e64dc9e43284e019093d0a9634b1048f\": container with ID starting with 6b34947dbaee53f404e0aa6070a1f595e64dc9e43284e019093d0a9634b1048f not found: ID does not exist" Oct 01 16:32:11 crc kubenswrapper[4726]: I1001 16:32:11.821980 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46a1d32f-b37b-48cb-880f-6cbb959a7b3a" path="/var/lib/kubelet/pods/46a1d32f-b37b-48cb-880f-6cbb959a7b3a/volumes" Oct 01 16:32:14 crc kubenswrapper[4726]: I1001 16:32:14.022681 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-xqs68" Oct 01 16:32:14 crc kubenswrapper[4726]: I1001 16:32:14.060138 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-xqs68" Oct 01 16:32:18 crc kubenswrapper[4726]: I1001 16:32:18.865041 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-mqdn2" Oct 01 16:32:18 crc kubenswrapper[4726]: I1001 16:32:18.951524 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7"] Oct 01 16:32:18 crc kubenswrapper[4726]: I1001 16:32:18.951754 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" podUID="0edf23ef-c2c4-4442-89ab-e736c6b989b5" containerName="frr-k8s-webhook-server" containerID="cri-o://c82a365f75111727835d96f54c55d87e42c823102a75b16d08f3f478beee622d" gracePeriod=10 Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.025942 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-xqs68" Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.261852 4726 generic.go:334] "Generic (PLEG): container finished" podID="0edf23ef-c2c4-4442-89ab-e736c6b989b5" containerID="c82a365f75111727835d96f54c55d87e42c823102a75b16d08f3f478beee622d" exitCode=0 Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.261913 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" event={"ID":"0edf23ef-c2c4-4442-89ab-e736c6b989b5","Type":"ContainerDied","Data":"c82a365f75111727835d96f54c55d87e42c823102a75b16d08f3f478beee622d"} Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.326826 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-kk7mh" Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.388252 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/controller-5d688f5ffc-gr6kx"] Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.388484 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/controller-5d688f5ffc-gr6kx" podUID="152cfa40-5a89-4715-825d-648e9f1c3d2f" containerName="controller" containerID="cri-o://359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99" gracePeriod=2 Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.388947 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/controller-5d688f5ffc-gr6kx" podUID="152cfa40-5a89-4715-825d-648e9f1c3d2f" containerName="kube-rbac-proxy" containerID="cri-o://5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73" gracePeriod=2 Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.402571 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/controller-5d688f5ffc-gr6kx"] Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.629339 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.647995 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpzp4\" (UniqueName: \"kubernetes.io/projected/0edf23ef-c2c4-4442-89ab-e736c6b989b5-kube-api-access-gpzp4\") pod \"0edf23ef-c2c4-4442-89ab-e736c6b989b5\" (UID: \"0edf23ef-c2c4-4442-89ab-e736c6b989b5\") " Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.659192 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0edf23ef-c2c4-4442-89ab-e736c6b989b5-kube-api-access-gpzp4" (OuterVolumeSpecName: "kube-api-access-gpzp4") pod "0edf23ef-c2c4-4442-89ab-e736c6b989b5" (UID: "0edf23ef-c2c4-4442-89ab-e736c6b989b5"). InnerVolumeSpecName "kube-api-access-gpzp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.736520 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.749293 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0edf23ef-c2c4-4442-89ab-e736c6b989b5-cert\") pod \"0edf23ef-c2c4-4442-89ab-e736c6b989b5\" (UID: \"0edf23ef-c2c4-4442-89ab-e736c6b989b5\") " Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.749541 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/152cfa40-5a89-4715-825d-648e9f1c3d2f-cert\") pod \"152cfa40-5a89-4715-825d-648e9f1c3d2f\" (UID: \"152cfa40-5a89-4715-825d-648e9f1c3d2f\") " Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.750298 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpzp4\" (UniqueName: \"kubernetes.io/projected/0edf23ef-c2c4-4442-89ab-e736c6b989b5-kube-api-access-gpzp4\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.753189 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0edf23ef-c2c4-4442-89ab-e736c6b989b5-cert" (OuterVolumeSpecName: "cert") pod "0edf23ef-c2c4-4442-89ab-e736c6b989b5" (UID: "0edf23ef-c2c4-4442-89ab-e736c6b989b5"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.754847 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/152cfa40-5a89-4715-825d-648e9f1c3d2f-cert" (OuterVolumeSpecName: "cert") pod "152cfa40-5a89-4715-825d-648e9f1c3d2f" (UID: "152cfa40-5a89-4715-825d-648e9f1c3d2f"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.851784 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/152cfa40-5a89-4715-825d-648e9f1c3d2f-metrics-certs\") pod \"152cfa40-5a89-4715-825d-648e9f1c3d2f\" (UID: \"152cfa40-5a89-4715-825d-648e9f1c3d2f\") " Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.851903 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rj5dv\" (UniqueName: \"kubernetes.io/projected/152cfa40-5a89-4715-825d-648e9f1c3d2f-kube-api-access-rj5dv\") pod \"152cfa40-5a89-4715-825d-648e9f1c3d2f\" (UID: \"152cfa40-5a89-4715-825d-648e9f1c3d2f\") " Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.855363 4726 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/152cfa40-5a89-4715-825d-648e9f1c3d2f-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.855393 4726 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0edf23ef-c2c4-4442-89ab-e736c6b989b5-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.856257 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/152cfa40-5a89-4715-825d-648e9f1c3d2f-kube-api-access-rj5dv" (OuterVolumeSpecName: "kube-api-access-rj5dv") pod "152cfa40-5a89-4715-825d-648e9f1c3d2f" (UID: "152cfa40-5a89-4715-825d-648e9f1c3d2f"). InnerVolumeSpecName "kube-api-access-rj5dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.859027 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/152cfa40-5a89-4715-825d-648e9f1c3d2f-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "152cfa40-5a89-4715-825d-648e9f1c3d2f" (UID: "152cfa40-5a89-4715-825d-648e9f1c3d2f"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.956588 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rj5dv\" (UniqueName: \"kubernetes.io/projected/152cfa40-5a89-4715-825d-648e9f1c3d2f-kube-api-access-rj5dv\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:19 crc kubenswrapper[4726]: I1001 16:32:19.956829 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/152cfa40-5a89-4715-825d-648e9f1c3d2f-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.272341 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.272336 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7" event={"ID":"0edf23ef-c2c4-4442-89ab-e736c6b989b5","Type":"ContainerDied","Data":"35e89045fafb920b2f096f35706772ba4511889304f17426a25b18009cd4c2dc"} Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.272513 4726 scope.go:117] "RemoveContainer" containerID="c82a365f75111727835d96f54c55d87e42c823102a75b16d08f3f478beee622d" Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.274660 4726 generic.go:334] "Generic (PLEG): container finished" podID="152cfa40-5a89-4715-825d-648e9f1c3d2f" containerID="5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73" exitCode=0 Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.274691 4726 generic.go:334] "Generic (PLEG): container finished" podID="152cfa40-5a89-4715-825d-648e9f1c3d2f" containerID="359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99" exitCode=0 Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.274722 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-gr6kx" Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.293139 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7"] Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.296088 4726 scope.go:117] "RemoveContainer" containerID="5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73" Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.306139 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-cpqd7"] Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.317339 4726 scope.go:117] "RemoveContainer" containerID="359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99" Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.338275 4726 scope.go:117] "RemoveContainer" containerID="5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73" Oct 01 16:32:20 crc kubenswrapper[4726]: E1001 16:32:20.338774 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73\": container with ID starting with 5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73 not found: ID does not exist" containerID="5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73" Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.338825 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73"} err="failed to get container status \"5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73\": rpc error: code = NotFound desc = could not find container \"5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73\": container with ID starting with 5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73 not found: ID does not exist" Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.338854 4726 scope.go:117] "RemoveContainer" containerID="359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99" Oct 01 16:32:20 crc kubenswrapper[4726]: E1001 16:32:20.339240 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99\": container with ID starting with 359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99 not found: ID does not exist" containerID="359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99" Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.339282 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99"} err="failed to get container status \"359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99\": rpc error: code = NotFound desc = could not find container \"359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99\": container with ID starting with 359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99 not found: ID does not exist" Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.339310 4726 scope.go:117] "RemoveContainer" containerID="5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73" Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.339655 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73"} err="failed to get container status \"5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73\": rpc error: code = NotFound desc = could not find container \"5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73\": container with ID starting with 5567a456a2b40bb405063f1b15ebace67af9c9fe252c4a60eea178301433cd73 not found: ID does not exist" Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.339676 4726 scope.go:117] "RemoveContainer" containerID="359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99" Oct 01 16:32:20 crc kubenswrapper[4726]: I1001 16:32:20.339905 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99"} err="failed to get container status \"359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99\": rpc error: code = NotFound desc = could not find container \"359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99\": container with ID starting with 359d45b990dc4103b20d81101a21e5021c2b01ece0d85653d3b5f30cb706ef99 not found: ID does not exist" Oct 01 16:32:21 crc kubenswrapper[4726]: I1001 16:32:21.819910 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0edf23ef-c2c4-4442-89ab-e736c6b989b5" path="/var/lib/kubelet/pods/0edf23ef-c2c4-4442-89ab-e736c6b989b5/volumes" Oct 01 16:32:21 crc kubenswrapper[4726]: I1001 16:32:21.821114 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="152cfa40-5a89-4715-825d-648e9f1c3d2f" path="/var/lib/kubelet/pods/152cfa40-5a89-4715-825d-648e9f1c3d2f/volumes" Oct 01 16:32:23 crc kubenswrapper[4726]: I1001 16:32:23.769383 4726 scope.go:117] "RemoveContainer" containerID="21e25a53c4de4535261a28d2622d716c55aa1ae4b3e304ee4cf0c9689471cb6c" Oct 01 16:32:23 crc kubenswrapper[4726]: I1001 16:32:23.806793 4726 scope.go:117] "RemoveContainer" containerID="1a181e670ed2692add3ac9550fd2f23b318e9905f01e3cc543a3dfe70da5d43a" Oct 01 16:32:23 crc kubenswrapper[4726]: I1001 16:32:23.840435 4726 scope.go:117] "RemoveContainer" containerID="b15081c4adf77afe809fbd87b422a48727f4320cbfa77d910c72a963c0b485f6" Oct 01 16:32:23 crc kubenswrapper[4726]: I1001 16:32:23.877419 4726 scope.go:117] "RemoveContainer" containerID="52c0e14d3038a74f8536411a7900dff96b2a9ae673439bf230ae87996a239754" Oct 01 16:32:23 crc kubenswrapper[4726]: I1001 16:32:23.896303 4726 scope.go:117] "RemoveContainer" containerID="abd2a7e13f99d37144d0076c20361976ce336847645daf450d24449bbadfe0ab" Oct 01 16:32:23 crc kubenswrapper[4726]: I1001 16:32:23.916025 4726 scope.go:117] "RemoveContainer" containerID="6c16210952121838cd718e7d1308ee55d2666e7f4be17b4431ce6f04bbf7195e" Oct 01 16:32:23 crc kubenswrapper[4726]: I1001 16:32:23.942913 4726 scope.go:117] "RemoveContainer" containerID="a1374a43ccb81a9a3f97068c3ca67cf90019219e90d280cff217e2b3c54782ef" Oct 01 16:32:23 crc kubenswrapper[4726]: I1001 16:32:23.990145 4726 scope.go:117] "RemoveContainer" containerID="33e1649f8c3dabd843a405caf4386c07f38d0c68d7380a48b0cbc7e7e679c3cc" Oct 01 16:32:24 crc kubenswrapper[4726]: I1001 16:32:24.020086 4726 scope.go:117] "RemoveContainer" containerID="d4481c40814f006b8e43ba761624c659b826b61501a2ac3cae57677c18dc7015" Oct 01 16:32:28 crc kubenswrapper[4726]: I1001 16:32:28.991788 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5fc7dbb978-2m6rr" Oct 01 16:32:29 crc kubenswrapper[4726]: I1001 16:32:29.096117 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2"] Oct 01 16:32:29 crc kubenswrapper[4726]: I1001 16:32:29.096326 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" podUID="f6b4559c-bbf0-4f5b-a3e6-d012136cf669" containerName="manager" containerID="cri-o://afa4ef815c35f4037fe2a0de09c92c79122752989b8ce5704f7f3e8878d5ed08" gracePeriod=10 Oct 01 16:32:29 crc kubenswrapper[4726]: I1001 16:32:29.377718 4726 generic.go:334] "Generic (PLEG): container finished" podID="f6b4559c-bbf0-4f5b-a3e6-d012136cf669" containerID="afa4ef815c35f4037fe2a0de09c92c79122752989b8ce5704f7f3e8878d5ed08" exitCode=0 Oct 01 16:32:29 crc kubenswrapper[4726]: I1001 16:32:29.377913 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" event={"ID":"f6b4559c-bbf0-4f5b-a3e6-d012136cf669","Type":"ContainerDied","Data":"afa4ef815c35f4037fe2a0de09c92c79122752989b8ce5704f7f3e8878d5ed08"} Oct 01 16:32:29 crc kubenswrapper[4726]: I1001 16:32:29.587624 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:32:29 crc kubenswrapper[4726]: I1001 16:32:29.659212 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-webhook-cert\") pod \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\" (UID: \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\") " Oct 01 16:32:29 crc kubenswrapper[4726]: I1001 16:32:29.659494 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-apiservice-cert\") pod \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\" (UID: \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\") " Oct 01 16:32:29 crc kubenswrapper[4726]: I1001 16:32:29.659538 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82bdc\" (UniqueName: \"kubernetes.io/projected/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-kube-api-access-82bdc\") pod \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\" (UID: \"f6b4559c-bbf0-4f5b-a3e6-d012136cf669\") " Oct 01 16:32:29 crc kubenswrapper[4726]: I1001 16:32:29.665016 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-kube-api-access-82bdc" (OuterVolumeSpecName: "kube-api-access-82bdc") pod "f6b4559c-bbf0-4f5b-a3e6-d012136cf669" (UID: "f6b4559c-bbf0-4f5b-a3e6-d012136cf669"). InnerVolumeSpecName "kube-api-access-82bdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:32:29 crc kubenswrapper[4726]: I1001 16:32:29.666342 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "f6b4559c-bbf0-4f5b-a3e6-d012136cf669" (UID: "f6b4559c-bbf0-4f5b-a3e6-d012136cf669"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:29 crc kubenswrapper[4726]: I1001 16:32:29.671231 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "f6b4559c-bbf0-4f5b-a3e6-d012136cf669" (UID: "f6b4559c-bbf0-4f5b-a3e6-d012136cf669"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:29 crc kubenswrapper[4726]: I1001 16:32:29.761998 4726 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:29 crc kubenswrapper[4726]: I1001 16:32:29.762035 4726 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:29 crc kubenswrapper[4726]: I1001 16:32:29.762071 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82bdc\" (UniqueName: \"kubernetes.io/projected/f6b4559c-bbf0-4f5b-a3e6-d012136cf669-kube-api-access-82bdc\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:30 crc kubenswrapper[4726]: I1001 16:32:30.388625 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" event={"ID":"f6b4559c-bbf0-4f5b-a3e6-d012136cf669","Type":"ContainerDied","Data":"39cad36fc9aa565fdbba64217ea21ee817e1fbc58b3533e407d45353c26e3ee9"} Oct 01 16:32:30 crc kubenswrapper[4726]: I1001 16:32:30.388683 4726 scope.go:117] "RemoveContainer" containerID="afa4ef815c35f4037fe2a0de09c92c79122752989b8ce5704f7f3e8878d5ed08" Oct 01 16:32:30 crc kubenswrapper[4726]: I1001 16:32:30.388710 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2" Oct 01 16:32:30 crc kubenswrapper[4726]: I1001 16:32:30.424991 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2"] Oct 01 16:32:30 crc kubenswrapper[4726]: I1001 16:32:30.433109 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d5479c8f-jc5f2"] Oct 01 16:32:31 crc kubenswrapper[4726]: I1001 16:32:31.825120 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6b4559c-bbf0-4f5b-a3e6-d012136cf669" path="/var/lib/kubelet/pods/f6b4559c-bbf0-4f5b-a3e6-d012136cf669/volumes" Oct 01 16:32:33 crc kubenswrapper[4726]: I1001 16:32:33.435297 4726 generic.go:334] "Generic (PLEG): container finished" podID="d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" containerID="dfbdfbb8e3069c25965a38d5fe75c4599f7ee394c95eeee223647dae2a8b7797" exitCode=0 Oct 01 16:32:33 crc kubenswrapper[4726]: I1001 16:32:33.435396 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" event={"ID":"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2","Type":"ContainerDied","Data":"dfbdfbb8e3069c25965a38d5fe75c4599f7ee394c95eeee223647dae2a8b7797"} Oct 01 16:32:34 crc kubenswrapper[4726]: I1001 16:32:34.927120 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.072344 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.072399 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-neutron-metadata-combined-ca-bundle\") pod \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.072467 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.072540 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-inventory\") pod \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.072573 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-repo-setup-combined-ca-bundle\") pod \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.072624 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-telemetry-combined-ca-bundle\") pod \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.072660 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-ovn-default-certs-0\") pod \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.072686 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-bootstrap-combined-ca-bundle\") pod \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.072712 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.072739 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2d6v\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-kube-api-access-k2d6v\") pod \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.072763 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-ovn-combined-ca-bundle\") pod \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.072832 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-libvirt-combined-ca-bundle\") pod \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.072864 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-ssh-key\") pod \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.072943 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-nova-combined-ca-bundle\") pod \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\" (UID: \"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2\") " Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.080447 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" (UID: "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.080520 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" (UID: "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.080685 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" (UID: "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.080711 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" (UID: "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.081336 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" (UID: "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.081385 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" (UID: "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.082293 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" (UID: "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.082610 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" (UID: "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.084286 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-kube-api-access-k2d6v" (OuterVolumeSpecName: "kube-api-access-k2d6v") pod "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" (UID: "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2"). InnerVolumeSpecName "kube-api-access-k2d6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.085185 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" (UID: "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.086481 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" (UID: "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.091263 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" (UID: "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.104606 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" (UID: "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.113423 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-inventory" (OuterVolumeSpecName: "inventory") pod "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" (UID: "d6a1971b-c5b2-4a12-90f2-d986c3ac37b2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.175689 4726 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.175725 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.175739 4726 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.175752 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.175764 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.175775 4726 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.175788 4726 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.175799 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.175809 4726 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.175822 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.175834 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2d6v\" (UniqueName: \"kubernetes.io/projected/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-kube-api-access-k2d6v\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.175846 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.175856 4726 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.175865 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6a1971b-c5b2-4a12-90f2-d986c3ac37b2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.459351 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" event={"ID":"d6a1971b-c5b2-4a12-90f2-d986c3ac37b2","Type":"ContainerDied","Data":"0897e1977b685b34baccb25a779ea9bd84c45f181abed3c372d3b9aafd9fe38d"} Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.459762 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0897e1977b685b34baccb25a779ea9bd84c45f181abed3c372d3b9aafd9fe38d" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.459424 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.551854 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9"] Oct 01 16:32:35 crc kubenswrapper[4726]: E1001 16:32:35.552275 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b4559c-bbf0-4f5b-a3e6-d012136cf669" containerName="manager" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.552293 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b4559c-bbf0-4f5b-a3e6-d012136cf669" containerName="manager" Oct 01 16:32:35 crc kubenswrapper[4726]: E1001 16:32:35.552312 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="152cfa40-5a89-4715-825d-648e9f1c3d2f" containerName="controller" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.552323 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="152cfa40-5a89-4715-825d-648e9f1c3d2f" containerName="controller" Oct 01 16:32:35 crc kubenswrapper[4726]: E1001 16:32:35.552334 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="152cfa40-5a89-4715-825d-648e9f1c3d2f" containerName="kube-rbac-proxy" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.552341 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="152cfa40-5a89-4715-825d-648e9f1c3d2f" containerName="kube-rbac-proxy" Oct 01 16:32:35 crc kubenswrapper[4726]: E1001 16:32:35.552375 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.552389 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 01 16:32:35 crc kubenswrapper[4726]: E1001 16:32:35.552424 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a1d32f-b37b-48cb-880f-6cbb959a7b3a" containerName="webhook-server" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.552433 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a1d32f-b37b-48cb-880f-6cbb959a7b3a" containerName="webhook-server" Oct 01 16:32:35 crc kubenswrapper[4726]: E1001 16:32:35.552450 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0edf23ef-c2c4-4442-89ab-e736c6b989b5" containerName="frr-k8s-webhook-server" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.552456 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0edf23ef-c2c4-4442-89ab-e736c6b989b5" containerName="frr-k8s-webhook-server" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.552625 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="152cfa40-5a89-4715-825d-648e9f1c3d2f" containerName="controller" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.552642 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a1971b-c5b2-4a12-90f2-d986c3ac37b2" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.552653 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="152cfa40-5a89-4715-825d-648e9f1c3d2f" containerName="kube-rbac-proxy" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.552662 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="46a1d32f-b37b-48cb-880f-6cbb959a7b3a" containerName="webhook-server" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.552674 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0edf23ef-c2c4-4442-89ab-e736c6b989b5" containerName="frr-k8s-webhook-server" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.552684 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b4559c-bbf0-4f5b-a3e6-d012136cf669" containerName="manager" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.553269 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.555601 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.556005 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.556679 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.556760 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.567265 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.569225 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9"] Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.688908 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.688991 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.689015 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.689035 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dp64\" (UniqueName: \"kubernetes.io/projected/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-kube-api-access-2dp64\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.689080 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.791397 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.791460 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.791481 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dp64\" (UniqueName: \"kubernetes.io/projected/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-kube-api-access-2dp64\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.791522 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.791624 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.792688 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.795722 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.796296 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.796787 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.814677 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dp64\" (UniqueName: \"kubernetes.io/projected/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-kube-api-access-2dp64\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ptlg9\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:35 crc kubenswrapper[4726]: I1001 16:32:35.870142 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:32:36 crc kubenswrapper[4726]: I1001 16:32:36.374859 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9"] Oct 01 16:32:36 crc kubenswrapper[4726]: I1001 16:32:36.476905 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" event={"ID":"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1","Type":"ContainerStarted","Data":"35c49fb10a4a00655a1c0d5f076dcfb86b5d7bf76bde201a206fbf26903caf32"} Oct 01 16:32:37 crc kubenswrapper[4726]: I1001 16:32:37.486670 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" event={"ID":"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1","Type":"ContainerStarted","Data":"a653c189d150de3871225c1db8cb88e05a5ce922d50e6216f78b452107c95f84"} Oct 01 16:32:37 crc kubenswrapper[4726]: I1001 16:32:37.503547 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" podStartSLOduration=1.987061267 podStartE2EDuration="2.503524269s" podCreationTimestamp="2025-10-01 16:32:35 +0000 UTC" firstStartedPulling="2025-10-01 16:32:36.3955496 +0000 UTC m=+2009.297102187" lastFinishedPulling="2025-10-01 16:32:36.912012612 +0000 UTC m=+2009.813565189" observedRunningTime="2025-10-01 16:32:37.502912032 +0000 UTC m=+2010.404464629" watchObservedRunningTime="2025-10-01 16:32:37.503524269 +0000 UTC m=+2010.405076856" Oct 01 16:32:53 crc kubenswrapper[4726]: I1001 16:32:53.413751 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:32:53 crc kubenswrapper[4726]: I1001 16:32:53.414476 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:33:23 crc kubenswrapper[4726]: I1001 16:33:23.414200 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:33:23 crc kubenswrapper[4726]: I1001 16:33:23.414762 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:33:39 crc kubenswrapper[4726]: I1001 16:33:39.082542 4726 generic.go:334] "Generic (PLEG): container finished" podID="4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1" containerID="a653c189d150de3871225c1db8cb88e05a5ce922d50e6216f78b452107c95f84" exitCode=0 Oct 01 16:33:39 crc kubenswrapper[4726]: I1001 16:33:39.082654 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" event={"ID":"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1","Type":"ContainerDied","Data":"a653c189d150de3871225c1db8cb88e05a5ce922d50e6216f78b452107c95f84"} Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.528319 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.621836 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ovncontroller-config-0\") pod \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.621920 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-inventory\") pod \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.622176 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ovn-combined-ca-bundle\") pod \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.622234 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dp64\" (UniqueName: \"kubernetes.io/projected/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-kube-api-access-2dp64\") pod \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.622311 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ssh-key\") pod \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\" (UID: \"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1\") " Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.628837 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1" (UID: "4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.629332 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-kube-api-access-2dp64" (OuterVolumeSpecName: "kube-api-access-2dp64") pod "4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1" (UID: "4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1"). InnerVolumeSpecName "kube-api-access-2dp64". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.656290 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1" (UID: "4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.659507 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1" (UID: "4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.670191 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-inventory" (OuterVolumeSpecName: "inventory") pod "4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1" (UID: "4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.725006 4726 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.725075 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.725089 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.725101 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dp64\" (UniqueName: \"kubernetes.io/projected/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-kube-api-access-2dp64\") on node \"crc\" DevicePath \"\"" Oct 01 16:33:40 crc kubenswrapper[4726]: I1001 16:33:40.725112 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.104031 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" event={"ID":"4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1","Type":"ContainerDied","Data":"35c49fb10a4a00655a1c0d5f076dcfb86b5d7bf76bde201a206fbf26903caf32"} Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.104335 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35c49fb10a4a00655a1c0d5f076dcfb86b5d7bf76bde201a206fbf26903caf32" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.104207 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ptlg9" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.194602 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6"] Oct 01 16:33:41 crc kubenswrapper[4726]: E1001 16:33:41.194932 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.194949 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.195167 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.195762 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.199956 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.200035 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.200070 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.200236 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.200593 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.200726 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.218892 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6"] Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.340778 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.340835 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.340863 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.341118 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.341146 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zml4t\" (UniqueName: \"kubernetes.io/projected/ef118af6-8d88-4bdd-bb43-c8532cd50f42-kube-api-access-zml4t\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.341170 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.443421 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.443524 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zml4t\" (UniqueName: \"kubernetes.io/projected/ef118af6-8d88-4bdd-bb43-c8532cd50f42-kube-api-access-zml4t\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.443558 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.444110 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.444146 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.444629 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.448772 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.448786 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.449971 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.450973 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.454993 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.460995 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zml4t\" (UniqueName: \"kubernetes.io/projected/ef118af6-8d88-4bdd-bb43-c8532cd50f42-kube-api-access-zml4t\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:41 crc kubenswrapper[4726]: I1001 16:33:41.525941 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:33:42 crc kubenswrapper[4726]: I1001 16:33:42.056611 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6"] Oct 01 16:33:42 crc kubenswrapper[4726]: I1001 16:33:42.114556 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" event={"ID":"ef118af6-8d88-4bdd-bb43-c8532cd50f42","Type":"ContainerStarted","Data":"8cdf520fba0a15947956b8bd09750eafa302988617e9d930a1c21ce1b2f9cd45"} Oct 01 16:33:44 crc kubenswrapper[4726]: I1001 16:33:44.137904 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" event={"ID":"ef118af6-8d88-4bdd-bb43-c8532cd50f42","Type":"ContainerStarted","Data":"74b57e2c0932b5273ac9e8f6b86ba561e708a673f29ef35151881665321b5011"} Oct 01 16:33:44 crc kubenswrapper[4726]: I1001 16:33:44.169075 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" podStartSLOduration=2.299382746 podStartE2EDuration="3.169020897s" podCreationTimestamp="2025-10-01 16:33:41 +0000 UTC" firstStartedPulling="2025-10-01 16:33:42.048165524 +0000 UTC m=+2074.949718101" lastFinishedPulling="2025-10-01 16:33:42.917803675 +0000 UTC m=+2075.819356252" observedRunningTime="2025-10-01 16:33:44.164657551 +0000 UTC m=+2077.066210138" watchObservedRunningTime="2025-10-01 16:33:44.169020897 +0000 UTC m=+2077.070573504" Oct 01 16:33:53 crc kubenswrapper[4726]: I1001 16:33:53.413980 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:33:53 crc kubenswrapper[4726]: I1001 16:33:53.415204 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:33:53 crc kubenswrapper[4726]: I1001 16:33:53.415290 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 16:33:53 crc kubenswrapper[4726]: I1001 16:33:53.416536 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ea6a11e2a728a74e08b777157550e1c40f0e83c9ab337602b7696ba954f563a"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:33:53 crc kubenswrapper[4726]: I1001 16:33:53.416615 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://8ea6a11e2a728a74e08b777157550e1c40f0e83c9ab337602b7696ba954f563a" gracePeriod=600 Oct 01 16:33:54 crc kubenswrapper[4726]: I1001 16:33:54.262180 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="8ea6a11e2a728a74e08b777157550e1c40f0e83c9ab337602b7696ba954f563a" exitCode=0 Oct 01 16:33:54 crc kubenswrapper[4726]: I1001 16:33:54.262338 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"8ea6a11e2a728a74e08b777157550e1c40f0e83c9ab337602b7696ba954f563a"} Oct 01 16:33:54 crc kubenswrapper[4726]: I1001 16:33:54.262812 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c"} Oct 01 16:33:54 crc kubenswrapper[4726]: I1001 16:33:54.262839 4726 scope.go:117] "RemoveContainer" containerID="513978cd8f62c6e6a40f5bed75842ab7a29231ad1c6b41ed77839804c0cfd840" Oct 01 16:34:31 crc kubenswrapper[4726]: I1001 16:34:31.620008 4726 generic.go:334] "Generic (PLEG): container finished" podID="ef118af6-8d88-4bdd-bb43-c8532cd50f42" containerID="74b57e2c0932b5273ac9e8f6b86ba561e708a673f29ef35151881665321b5011" exitCode=0 Oct 01 16:34:31 crc kubenswrapper[4726]: I1001 16:34:31.620132 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" event={"ID":"ef118af6-8d88-4bdd-bb43-c8532cd50f42","Type":"ContainerDied","Data":"74b57e2c0932b5273ac9e8f6b86ba561e708a673f29ef35151881665321b5011"} Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.059346 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.158311 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-neutron-metadata-combined-ca-bundle\") pod \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.158431 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-ssh-key\") pod \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.158521 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-nova-metadata-neutron-config-0\") pod \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.158617 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zml4t\" (UniqueName: \"kubernetes.io/projected/ef118af6-8d88-4bdd-bb43-c8532cd50f42-kube-api-access-zml4t\") pod \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.158680 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-inventory\") pod \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.158725 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-neutron-ovn-metadata-agent-neutron-config-0\") pod \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\" (UID: \"ef118af6-8d88-4bdd-bb43-c8532cd50f42\") " Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.164716 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "ef118af6-8d88-4bdd-bb43-c8532cd50f42" (UID: "ef118af6-8d88-4bdd-bb43-c8532cd50f42"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.165062 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef118af6-8d88-4bdd-bb43-c8532cd50f42-kube-api-access-zml4t" (OuterVolumeSpecName: "kube-api-access-zml4t") pod "ef118af6-8d88-4bdd-bb43-c8532cd50f42" (UID: "ef118af6-8d88-4bdd-bb43-c8532cd50f42"). InnerVolumeSpecName "kube-api-access-zml4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.189762 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "ef118af6-8d88-4bdd-bb43-c8532cd50f42" (UID: "ef118af6-8d88-4bdd-bb43-c8532cd50f42"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.193258 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ef118af6-8d88-4bdd-bb43-c8532cd50f42" (UID: "ef118af6-8d88-4bdd-bb43-c8532cd50f42"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.195597 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-inventory" (OuterVolumeSpecName: "inventory") pod "ef118af6-8d88-4bdd-bb43-c8532cd50f42" (UID: "ef118af6-8d88-4bdd-bb43-c8532cd50f42"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.197591 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "ef118af6-8d88-4bdd-bb43-c8532cd50f42" (UID: "ef118af6-8d88-4bdd-bb43-c8532cd50f42"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.262404 4726 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.262476 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.262504 4726 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.262534 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zml4t\" (UniqueName: \"kubernetes.io/projected/ef118af6-8d88-4bdd-bb43-c8532cd50f42-kube-api-access-zml4t\") on node \"crc\" DevicePath \"\"" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.262561 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.262586 4726 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ef118af6-8d88-4bdd-bb43-c8532cd50f42-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.640499 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" event={"ID":"ef118af6-8d88-4bdd-bb43-c8532cd50f42","Type":"ContainerDied","Data":"8cdf520fba0a15947956b8bd09750eafa302988617e9d930a1c21ce1b2f9cd45"} Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.640538 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cdf520fba0a15947956b8bd09750eafa302988617e9d930a1c21ce1b2f9cd45" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.640588 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.780509 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg"] Oct 01 16:34:33 crc kubenswrapper[4726]: E1001 16:34:33.781414 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef118af6-8d88-4bdd-bb43-c8532cd50f42" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.781433 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef118af6-8d88-4bdd-bb43-c8532cd50f42" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.781682 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef118af6-8d88-4bdd-bb43-c8532cd50f42" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.782561 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.793354 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.793457 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.793934 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.797802 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.797942 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.821855 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg"] Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.875017 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.875098 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.875150 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm8lw\" (UniqueName: \"kubernetes.io/projected/a66fccbb-d629-4290-a469-066027643cfa-kube-api-access-dm8lw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.875287 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.875480 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.977619 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.977673 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.977727 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm8lw\" (UniqueName: \"kubernetes.io/projected/a66fccbb-d629-4290-a469-066027643cfa-kube-api-access-dm8lw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.977749 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.977808 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.983067 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.984963 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.988830 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.989237 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:33 crc kubenswrapper[4726]: I1001 16:34:33.997830 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm8lw\" (UniqueName: \"kubernetes.io/projected/a66fccbb-d629-4290-a469-066027643cfa-kube-api-access-dm8lw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-svtfg\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:34 crc kubenswrapper[4726]: I1001 16:34:34.107686 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:34:34 crc kubenswrapper[4726]: I1001 16:34:34.613683 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg"] Oct 01 16:34:34 crc kubenswrapper[4726]: W1001 16:34:34.622565 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda66fccbb_d629_4290_a469_066027643cfa.slice/crio-e0bbe0fe0d459b52c1bfda2a7a7d600699bb7d3e48c52928b4d50909bfcb5b15 WatchSource:0}: Error finding container e0bbe0fe0d459b52c1bfda2a7a7d600699bb7d3e48c52928b4d50909bfcb5b15: Status 404 returned error can't find the container with id e0bbe0fe0d459b52c1bfda2a7a7d600699bb7d3e48c52928b4d50909bfcb5b15 Oct 01 16:34:34 crc kubenswrapper[4726]: I1001 16:34:34.651342 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" event={"ID":"a66fccbb-d629-4290-a469-066027643cfa","Type":"ContainerStarted","Data":"e0bbe0fe0d459b52c1bfda2a7a7d600699bb7d3e48c52928b4d50909bfcb5b15"} Oct 01 16:34:35 crc kubenswrapper[4726]: I1001 16:34:35.664018 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" event={"ID":"a66fccbb-d629-4290-a469-066027643cfa","Type":"ContainerStarted","Data":"69980698b0be0d842211ac5ab670385b30af0f68306e71515e421952101a311c"} Oct 01 16:34:35 crc kubenswrapper[4726]: I1001 16:34:35.688160 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" podStartSLOduration=2.138555745 podStartE2EDuration="2.688135498s" podCreationTimestamp="2025-10-01 16:34:33 +0000 UTC" firstStartedPulling="2025-10-01 16:34:34.625204494 +0000 UTC m=+2127.526757071" lastFinishedPulling="2025-10-01 16:34:35.174784217 +0000 UTC m=+2128.076336824" observedRunningTime="2025-10-01 16:34:35.677888331 +0000 UTC m=+2128.579440938" watchObservedRunningTime="2025-10-01 16:34:35.688135498 +0000 UTC m=+2128.589688075" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.043542 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-87hl7"] Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.049351 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.054694 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-87hl7"] Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.164144 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7db561a3-0344-4346-8276-cc7770b5efac-utilities\") pod \"certified-operators-87hl7\" (UID: \"7db561a3-0344-4346-8276-cc7770b5efac\") " pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.164558 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7db561a3-0344-4346-8276-cc7770b5efac-catalog-content\") pod \"certified-operators-87hl7\" (UID: \"7db561a3-0344-4346-8276-cc7770b5efac\") " pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.164648 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgmnt\" (UniqueName: \"kubernetes.io/projected/7db561a3-0344-4346-8276-cc7770b5efac-kube-api-access-hgmnt\") pod \"certified-operators-87hl7\" (UID: \"7db561a3-0344-4346-8276-cc7770b5efac\") " pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.241842 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zg78g"] Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.243776 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.266364 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgmnt\" (UniqueName: \"kubernetes.io/projected/7db561a3-0344-4346-8276-cc7770b5efac-kube-api-access-hgmnt\") pod \"certified-operators-87hl7\" (UID: \"7db561a3-0344-4346-8276-cc7770b5efac\") " pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.266522 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7db561a3-0344-4346-8276-cc7770b5efac-utilities\") pod \"certified-operators-87hl7\" (UID: \"7db561a3-0344-4346-8276-cc7770b5efac\") " pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.266563 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7db561a3-0344-4346-8276-cc7770b5efac-catalog-content\") pod \"certified-operators-87hl7\" (UID: \"7db561a3-0344-4346-8276-cc7770b5efac\") " pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.267601 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7db561a3-0344-4346-8276-cc7770b5efac-catalog-content\") pod \"certified-operators-87hl7\" (UID: \"7db561a3-0344-4346-8276-cc7770b5efac\") " pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.267648 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7db561a3-0344-4346-8276-cc7770b5efac-utilities\") pod \"certified-operators-87hl7\" (UID: \"7db561a3-0344-4346-8276-cc7770b5efac\") " pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.275629 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zg78g"] Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.295524 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgmnt\" (UniqueName: \"kubernetes.io/projected/7db561a3-0344-4346-8276-cc7770b5efac-kube-api-access-hgmnt\") pod \"certified-operators-87hl7\" (UID: \"7db561a3-0344-4346-8276-cc7770b5efac\") " pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.368448 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-utilities\") pod \"redhat-marketplace-zg78g\" (UID: \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\") " pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.368801 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2l5m\" (UniqueName: \"kubernetes.io/projected/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-kube-api-access-n2l5m\") pod \"redhat-marketplace-zg78g\" (UID: \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\") " pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.369121 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-catalog-content\") pod \"redhat-marketplace-zg78g\" (UID: \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\") " pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.369878 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.472271 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-utilities\") pod \"redhat-marketplace-zg78g\" (UID: \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\") " pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.472407 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2l5m\" (UniqueName: \"kubernetes.io/projected/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-kube-api-access-n2l5m\") pod \"redhat-marketplace-zg78g\" (UID: \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\") " pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.472494 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-catalog-content\") pod \"redhat-marketplace-zg78g\" (UID: \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\") " pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.473110 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-catalog-content\") pod \"redhat-marketplace-zg78g\" (UID: \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\") " pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.473714 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-utilities\") pod \"redhat-marketplace-zg78g\" (UID: \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\") " pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.532234 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2l5m\" (UniqueName: \"kubernetes.io/projected/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-kube-api-access-n2l5m\") pod \"redhat-marketplace-zg78g\" (UID: \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\") " pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.570781 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:34:53 crc kubenswrapper[4726]: I1001 16:34:53.932263 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-87hl7"] Oct 01 16:34:54 crc kubenswrapper[4726]: I1001 16:34:54.132611 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zg78g"] Oct 01 16:34:54 crc kubenswrapper[4726]: W1001 16:34:54.173144 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ea6455f_dafb_46d1_94e4_8ed5ff94ab97.slice/crio-ca01f5ea200b530a5393dddd4ad91f3bc52b2226337dbdbc55da41b12cf0eccd WatchSource:0}: Error finding container ca01f5ea200b530a5393dddd4ad91f3bc52b2226337dbdbc55da41b12cf0eccd: Status 404 returned error can't find the container with id ca01f5ea200b530a5393dddd4ad91f3bc52b2226337dbdbc55da41b12cf0eccd Oct 01 16:34:54 crc kubenswrapper[4726]: I1001 16:34:54.875151 4726 generic.go:334] "Generic (PLEG): container finished" podID="7db561a3-0344-4346-8276-cc7770b5efac" containerID="75a77463bbd47bdd174a039d77e1a5fc45194cdc9b4a3223e6c8370e1c3ff1f5" exitCode=0 Oct 01 16:34:54 crc kubenswrapper[4726]: I1001 16:34:54.875344 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87hl7" event={"ID":"7db561a3-0344-4346-8276-cc7770b5efac","Type":"ContainerDied","Data":"75a77463bbd47bdd174a039d77e1a5fc45194cdc9b4a3223e6c8370e1c3ff1f5"} Oct 01 16:34:54 crc kubenswrapper[4726]: I1001 16:34:54.875479 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87hl7" event={"ID":"7db561a3-0344-4346-8276-cc7770b5efac","Type":"ContainerStarted","Data":"8c85b2c4ca0e9f14a691a84ffaa958a7b015ccdbfe1b91a57bf2102e77061c7d"} Oct 01 16:34:54 crc kubenswrapper[4726]: I1001 16:34:54.880956 4726 generic.go:334] "Generic (PLEG): container finished" podID="1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" containerID="9cde1131f2873aa349ecfbc06a6b1624388256e73e52a0d70f08df4bc3152ba3" exitCode=0 Oct 01 16:34:54 crc kubenswrapper[4726]: I1001 16:34:54.881031 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg78g" event={"ID":"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97","Type":"ContainerDied","Data":"9cde1131f2873aa349ecfbc06a6b1624388256e73e52a0d70f08df4bc3152ba3"} Oct 01 16:34:54 crc kubenswrapper[4726]: I1001 16:34:54.881248 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg78g" event={"ID":"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97","Type":"ContainerStarted","Data":"ca01f5ea200b530a5393dddd4ad91f3bc52b2226337dbdbc55da41b12cf0eccd"} Oct 01 16:34:57 crc kubenswrapper[4726]: I1001 16:34:57.914194 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87hl7" event={"ID":"7db561a3-0344-4346-8276-cc7770b5efac","Type":"ContainerStarted","Data":"e5766b2bfaf8ab5f8f498f1f1b733043dc22afcac167e597c8b1c88809e15455"} Oct 01 16:34:57 crc kubenswrapper[4726]: I1001 16:34:57.916992 4726 generic.go:334] "Generic (PLEG): container finished" podID="1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" containerID="40c05fb4bd4af064738ef7761cd10efde43a4b81876477af18dc1f5b4edf97d4" exitCode=0 Oct 01 16:34:57 crc kubenswrapper[4726]: I1001 16:34:57.917079 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg78g" event={"ID":"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97","Type":"ContainerDied","Data":"40c05fb4bd4af064738ef7761cd10efde43a4b81876477af18dc1f5b4edf97d4"} Oct 01 16:35:01 crc kubenswrapper[4726]: I1001 16:35:01.956696 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg78g" event={"ID":"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97","Type":"ContainerStarted","Data":"16b3195928b8853a5e5fac3c3642e8ce340ea98fa08614e8fe07bd3a6bf89226"} Oct 01 16:35:01 crc kubenswrapper[4726]: I1001 16:35:01.959610 4726 generic.go:334] "Generic (PLEG): container finished" podID="7db561a3-0344-4346-8276-cc7770b5efac" containerID="e5766b2bfaf8ab5f8f498f1f1b733043dc22afcac167e597c8b1c88809e15455" exitCode=0 Oct 01 16:35:01 crc kubenswrapper[4726]: I1001 16:35:01.959649 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87hl7" event={"ID":"7db561a3-0344-4346-8276-cc7770b5efac","Type":"ContainerDied","Data":"e5766b2bfaf8ab5f8f498f1f1b733043dc22afcac167e597c8b1c88809e15455"} Oct 01 16:35:01 crc kubenswrapper[4726]: I1001 16:35:01.983997 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zg78g" podStartSLOduration=3.107603841 podStartE2EDuration="8.983974365s" podCreationTimestamp="2025-10-01 16:34:53 +0000 UTC" firstStartedPulling="2025-10-01 16:34:54.883983738 +0000 UTC m=+2147.785536315" lastFinishedPulling="2025-10-01 16:35:00.760354252 +0000 UTC m=+2153.661906839" observedRunningTime="2025-10-01 16:35:01.975415797 +0000 UTC m=+2154.876968394" watchObservedRunningTime="2025-10-01 16:35:01.983974365 +0000 UTC m=+2154.885526942" Oct 01 16:35:02 crc kubenswrapper[4726]: I1001 16:35:02.970318 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87hl7" event={"ID":"7db561a3-0344-4346-8276-cc7770b5efac","Type":"ContainerStarted","Data":"8d9f599bb19a6c3d63f1653d0ecdf4e7035ae0ecb973b6d045302bd139b2f761"} Oct 01 16:35:03 crc kubenswrapper[4726]: I1001 16:35:03.571596 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:35:03 crc kubenswrapper[4726]: I1001 16:35:03.571658 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:35:03 crc kubenswrapper[4726]: I1001 16:35:03.620524 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:35:04 crc kubenswrapper[4726]: I1001 16:35:04.009020 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-87hl7" podStartSLOduration=3.38006843 podStartE2EDuration="11.009003411s" podCreationTimestamp="2025-10-01 16:34:53 +0000 UTC" firstStartedPulling="2025-10-01 16:34:54.878729496 +0000 UTC m=+2147.780282073" lastFinishedPulling="2025-10-01 16:35:02.507664477 +0000 UTC m=+2155.409217054" observedRunningTime="2025-10-01 16:35:04.001713879 +0000 UTC m=+2156.903266466" watchObservedRunningTime="2025-10-01 16:35:04.009003411 +0000 UTC m=+2156.910555988" Oct 01 16:35:13 crc kubenswrapper[4726]: I1001 16:35:13.370551 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:35:13 crc kubenswrapper[4726]: I1001 16:35:13.371357 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:35:13 crc kubenswrapper[4726]: I1001 16:35:13.426987 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:35:13 crc kubenswrapper[4726]: I1001 16:35:13.635741 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:35:13 crc kubenswrapper[4726]: I1001 16:35:13.700608 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zg78g"] Oct 01 16:35:14 crc kubenswrapper[4726]: I1001 16:35:14.068123 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zg78g" podUID="1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" containerName="registry-server" containerID="cri-o://16b3195928b8853a5e5fac3c3642e8ce340ea98fa08614e8fe07bd3a6bf89226" gracePeriod=2 Oct 01 16:35:14 crc kubenswrapper[4726]: I1001 16:35:14.148192 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:35:14 crc kubenswrapper[4726]: I1001 16:35:14.579640 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:35:14 crc kubenswrapper[4726]: I1001 16:35:14.663397 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-utilities\") pod \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\" (UID: \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\") " Oct 01 16:35:14 crc kubenswrapper[4726]: I1001 16:35:14.663659 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2l5m\" (UniqueName: \"kubernetes.io/projected/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-kube-api-access-n2l5m\") pod \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\" (UID: \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\") " Oct 01 16:35:14 crc kubenswrapper[4726]: I1001 16:35:14.663769 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-catalog-content\") pod \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\" (UID: \"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97\") " Oct 01 16:35:14 crc kubenswrapper[4726]: I1001 16:35:14.664328 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-utilities" (OuterVolumeSpecName: "utilities") pod "1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" (UID: "1ea6455f-dafb-46d1-94e4-8ed5ff94ab97"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:35:14 crc kubenswrapper[4726]: I1001 16:35:14.671703 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-kube-api-access-n2l5m" (OuterVolumeSpecName: "kube-api-access-n2l5m") pod "1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" (UID: "1ea6455f-dafb-46d1-94e4-8ed5ff94ab97"). InnerVolumeSpecName "kube-api-access-n2l5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:35:14 crc kubenswrapper[4726]: I1001 16:35:14.681465 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" (UID: "1ea6455f-dafb-46d1-94e4-8ed5ff94ab97"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:35:14 crc kubenswrapper[4726]: I1001 16:35:14.765979 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:35:14 crc kubenswrapper[4726]: I1001 16:35:14.766011 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2l5m\" (UniqueName: \"kubernetes.io/projected/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-kube-api-access-n2l5m\") on node \"crc\" DevicePath \"\"" Oct 01 16:35:14 crc kubenswrapper[4726]: I1001 16:35:14.766020 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.085422 4726 generic.go:334] "Generic (PLEG): container finished" podID="1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" containerID="16b3195928b8853a5e5fac3c3642e8ce340ea98fa08614e8fe07bd3a6bf89226" exitCode=0 Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.085531 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zg78g" Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.085531 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg78g" event={"ID":"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97","Type":"ContainerDied","Data":"16b3195928b8853a5e5fac3c3642e8ce340ea98fa08614e8fe07bd3a6bf89226"} Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.086107 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg78g" event={"ID":"1ea6455f-dafb-46d1-94e4-8ed5ff94ab97","Type":"ContainerDied","Data":"ca01f5ea200b530a5393dddd4ad91f3bc52b2226337dbdbc55da41b12cf0eccd"} Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.086168 4726 scope.go:117] "RemoveContainer" containerID="16b3195928b8853a5e5fac3c3642e8ce340ea98fa08614e8fe07bd3a6bf89226" Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.114212 4726 scope.go:117] "RemoveContainer" containerID="40c05fb4bd4af064738ef7761cd10efde43a4b81876477af18dc1f5b4edf97d4" Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.139945 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zg78g"] Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.154440 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zg78g"] Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.162803 4726 scope.go:117] "RemoveContainer" containerID="9cde1131f2873aa349ecfbc06a6b1624388256e73e52a0d70f08df4bc3152ba3" Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.207210 4726 scope.go:117] "RemoveContainer" containerID="16b3195928b8853a5e5fac3c3642e8ce340ea98fa08614e8fe07bd3a6bf89226" Oct 01 16:35:15 crc kubenswrapper[4726]: E1001 16:35:15.208266 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16b3195928b8853a5e5fac3c3642e8ce340ea98fa08614e8fe07bd3a6bf89226\": container with ID starting with 16b3195928b8853a5e5fac3c3642e8ce340ea98fa08614e8fe07bd3a6bf89226 not found: ID does not exist" containerID="16b3195928b8853a5e5fac3c3642e8ce340ea98fa08614e8fe07bd3a6bf89226" Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.208305 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16b3195928b8853a5e5fac3c3642e8ce340ea98fa08614e8fe07bd3a6bf89226"} err="failed to get container status \"16b3195928b8853a5e5fac3c3642e8ce340ea98fa08614e8fe07bd3a6bf89226\": rpc error: code = NotFound desc = could not find container \"16b3195928b8853a5e5fac3c3642e8ce340ea98fa08614e8fe07bd3a6bf89226\": container with ID starting with 16b3195928b8853a5e5fac3c3642e8ce340ea98fa08614e8fe07bd3a6bf89226 not found: ID does not exist" Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.208333 4726 scope.go:117] "RemoveContainer" containerID="40c05fb4bd4af064738ef7761cd10efde43a4b81876477af18dc1f5b4edf97d4" Oct 01 16:35:15 crc kubenswrapper[4726]: E1001 16:35:15.208827 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40c05fb4bd4af064738ef7761cd10efde43a4b81876477af18dc1f5b4edf97d4\": container with ID starting with 40c05fb4bd4af064738ef7761cd10efde43a4b81876477af18dc1f5b4edf97d4 not found: ID does not exist" containerID="40c05fb4bd4af064738ef7761cd10efde43a4b81876477af18dc1f5b4edf97d4" Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.208853 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40c05fb4bd4af064738ef7761cd10efde43a4b81876477af18dc1f5b4edf97d4"} err="failed to get container status \"40c05fb4bd4af064738ef7761cd10efde43a4b81876477af18dc1f5b4edf97d4\": rpc error: code = NotFound desc = could not find container \"40c05fb4bd4af064738ef7761cd10efde43a4b81876477af18dc1f5b4edf97d4\": container with ID starting with 40c05fb4bd4af064738ef7761cd10efde43a4b81876477af18dc1f5b4edf97d4 not found: ID does not exist" Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.208867 4726 scope.go:117] "RemoveContainer" containerID="9cde1131f2873aa349ecfbc06a6b1624388256e73e52a0d70f08df4bc3152ba3" Oct 01 16:35:15 crc kubenswrapper[4726]: E1001 16:35:15.209206 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cde1131f2873aa349ecfbc06a6b1624388256e73e52a0d70f08df4bc3152ba3\": container with ID starting with 9cde1131f2873aa349ecfbc06a6b1624388256e73e52a0d70f08df4bc3152ba3 not found: ID does not exist" containerID="9cde1131f2873aa349ecfbc06a6b1624388256e73e52a0d70f08df4bc3152ba3" Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.209229 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cde1131f2873aa349ecfbc06a6b1624388256e73e52a0d70f08df4bc3152ba3"} err="failed to get container status \"9cde1131f2873aa349ecfbc06a6b1624388256e73e52a0d70f08df4bc3152ba3\": rpc error: code = NotFound desc = could not find container \"9cde1131f2873aa349ecfbc06a6b1624388256e73e52a0d70f08df4bc3152ba3\": container with ID starting with 9cde1131f2873aa349ecfbc06a6b1624388256e73e52a0d70f08df4bc3152ba3 not found: ID does not exist" Oct 01 16:35:15 crc kubenswrapper[4726]: I1001 16:35:15.824950 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" path="/var/lib/kubelet/pods/1ea6455f-dafb-46d1-94e4-8ed5ff94ab97/volumes" Oct 01 16:35:16 crc kubenswrapper[4726]: I1001 16:35:16.474896 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-87hl7"] Oct 01 16:35:16 crc kubenswrapper[4726]: I1001 16:35:16.475765 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-87hl7" podUID="7db561a3-0344-4346-8276-cc7770b5efac" containerName="registry-server" containerID="cri-o://8d9f599bb19a6c3d63f1653d0ecdf4e7035ae0ecb973b6d045302bd139b2f761" gracePeriod=2 Oct 01 16:35:16 crc kubenswrapper[4726]: I1001 16:35:16.970001 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.133309 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7db561a3-0344-4346-8276-cc7770b5efac-utilities\") pod \"7db561a3-0344-4346-8276-cc7770b5efac\" (UID: \"7db561a3-0344-4346-8276-cc7770b5efac\") " Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.133672 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgmnt\" (UniqueName: \"kubernetes.io/projected/7db561a3-0344-4346-8276-cc7770b5efac-kube-api-access-hgmnt\") pod \"7db561a3-0344-4346-8276-cc7770b5efac\" (UID: \"7db561a3-0344-4346-8276-cc7770b5efac\") " Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.133706 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7db561a3-0344-4346-8276-cc7770b5efac-catalog-content\") pod \"7db561a3-0344-4346-8276-cc7770b5efac\" (UID: \"7db561a3-0344-4346-8276-cc7770b5efac\") " Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.140676 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7db561a3-0344-4346-8276-cc7770b5efac-kube-api-access-hgmnt" (OuterVolumeSpecName: "kube-api-access-hgmnt") pod "7db561a3-0344-4346-8276-cc7770b5efac" (UID: "7db561a3-0344-4346-8276-cc7770b5efac"). InnerVolumeSpecName "kube-api-access-hgmnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.150318 4726 generic.go:334] "Generic (PLEG): container finished" podID="7db561a3-0344-4346-8276-cc7770b5efac" containerID="8d9f599bb19a6c3d63f1653d0ecdf4e7035ae0ecb973b6d045302bd139b2f761" exitCode=0 Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.150396 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87hl7" event={"ID":"7db561a3-0344-4346-8276-cc7770b5efac","Type":"ContainerDied","Data":"8d9f599bb19a6c3d63f1653d0ecdf4e7035ae0ecb973b6d045302bd139b2f761"} Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.150405 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-87hl7" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.150450 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87hl7" event={"ID":"7db561a3-0344-4346-8276-cc7770b5efac","Type":"ContainerDied","Data":"8c85b2c4ca0e9f14a691a84ffaa958a7b015ccdbfe1b91a57bf2102e77061c7d"} Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.150483 4726 scope.go:117] "RemoveContainer" containerID="8d9f599bb19a6c3d63f1653d0ecdf4e7035ae0ecb973b6d045302bd139b2f761" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.155779 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7db561a3-0344-4346-8276-cc7770b5efac-utilities" (OuterVolumeSpecName: "utilities") pod "7db561a3-0344-4346-8276-cc7770b5efac" (UID: "7db561a3-0344-4346-8276-cc7770b5efac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.236901 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7db561a3-0344-4346-8276-cc7770b5efac-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.236957 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgmnt\" (UniqueName: \"kubernetes.io/projected/7db561a3-0344-4346-8276-cc7770b5efac-kube-api-access-hgmnt\") on node \"crc\" DevicePath \"\"" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.292268 4726 scope.go:117] "RemoveContainer" containerID="e5766b2bfaf8ab5f8f498f1f1b733043dc22afcac167e597c8b1c88809e15455" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.329554 4726 scope.go:117] "RemoveContainer" containerID="75a77463bbd47bdd174a039d77e1a5fc45194cdc9b4a3223e6c8370e1c3ff1f5" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.351963 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7db561a3-0344-4346-8276-cc7770b5efac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7db561a3-0344-4346-8276-cc7770b5efac" (UID: "7db561a3-0344-4346-8276-cc7770b5efac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.384997 4726 scope.go:117] "RemoveContainer" containerID="8d9f599bb19a6c3d63f1653d0ecdf4e7035ae0ecb973b6d045302bd139b2f761" Oct 01 16:35:17 crc kubenswrapper[4726]: E1001 16:35:17.385729 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d9f599bb19a6c3d63f1653d0ecdf4e7035ae0ecb973b6d045302bd139b2f761\": container with ID starting with 8d9f599bb19a6c3d63f1653d0ecdf4e7035ae0ecb973b6d045302bd139b2f761 not found: ID does not exist" containerID="8d9f599bb19a6c3d63f1653d0ecdf4e7035ae0ecb973b6d045302bd139b2f761" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.385820 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d9f599bb19a6c3d63f1653d0ecdf4e7035ae0ecb973b6d045302bd139b2f761"} err="failed to get container status \"8d9f599bb19a6c3d63f1653d0ecdf4e7035ae0ecb973b6d045302bd139b2f761\": rpc error: code = NotFound desc = could not find container \"8d9f599bb19a6c3d63f1653d0ecdf4e7035ae0ecb973b6d045302bd139b2f761\": container with ID starting with 8d9f599bb19a6c3d63f1653d0ecdf4e7035ae0ecb973b6d045302bd139b2f761 not found: ID does not exist" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.385849 4726 scope.go:117] "RemoveContainer" containerID="e5766b2bfaf8ab5f8f498f1f1b733043dc22afcac167e597c8b1c88809e15455" Oct 01 16:35:17 crc kubenswrapper[4726]: E1001 16:35:17.386454 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5766b2bfaf8ab5f8f498f1f1b733043dc22afcac167e597c8b1c88809e15455\": container with ID starting with e5766b2bfaf8ab5f8f498f1f1b733043dc22afcac167e597c8b1c88809e15455 not found: ID does not exist" containerID="e5766b2bfaf8ab5f8f498f1f1b733043dc22afcac167e597c8b1c88809e15455" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.386499 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5766b2bfaf8ab5f8f498f1f1b733043dc22afcac167e597c8b1c88809e15455"} err="failed to get container status \"e5766b2bfaf8ab5f8f498f1f1b733043dc22afcac167e597c8b1c88809e15455\": rpc error: code = NotFound desc = could not find container \"e5766b2bfaf8ab5f8f498f1f1b733043dc22afcac167e597c8b1c88809e15455\": container with ID starting with e5766b2bfaf8ab5f8f498f1f1b733043dc22afcac167e597c8b1c88809e15455 not found: ID does not exist" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.386529 4726 scope.go:117] "RemoveContainer" containerID="75a77463bbd47bdd174a039d77e1a5fc45194cdc9b4a3223e6c8370e1c3ff1f5" Oct 01 16:35:17 crc kubenswrapper[4726]: E1001 16:35:17.386939 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75a77463bbd47bdd174a039d77e1a5fc45194cdc9b4a3223e6c8370e1c3ff1f5\": container with ID starting with 75a77463bbd47bdd174a039d77e1a5fc45194cdc9b4a3223e6c8370e1c3ff1f5 not found: ID does not exist" containerID="75a77463bbd47bdd174a039d77e1a5fc45194cdc9b4a3223e6c8370e1c3ff1f5" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.386973 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75a77463bbd47bdd174a039d77e1a5fc45194cdc9b4a3223e6c8370e1c3ff1f5"} err="failed to get container status \"75a77463bbd47bdd174a039d77e1a5fc45194cdc9b4a3223e6c8370e1c3ff1f5\": rpc error: code = NotFound desc = could not find container \"75a77463bbd47bdd174a039d77e1a5fc45194cdc9b4a3223e6c8370e1c3ff1f5\": container with ID starting with 75a77463bbd47bdd174a039d77e1a5fc45194cdc9b4a3223e6c8370e1c3ff1f5 not found: ID does not exist" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.440391 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7db561a3-0344-4346-8276-cc7770b5efac-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.499963 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-87hl7"] Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.508648 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-87hl7"] Oct 01 16:35:17 crc kubenswrapper[4726]: I1001 16:35:17.823633 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7db561a3-0344-4346-8276-cc7770b5efac" path="/var/lib/kubelet/pods/7db561a3-0344-4346-8276-cc7770b5efac/volumes" Oct 01 16:35:53 crc kubenswrapper[4726]: I1001 16:35:53.413454 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:35:53 crc kubenswrapper[4726]: I1001 16:35:53.414175 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:36:23 crc kubenswrapper[4726]: I1001 16:36:23.413899 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:36:23 crc kubenswrapper[4726]: I1001 16:36:23.414564 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.463127 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-px6g2"] Oct 01 16:36:34 crc kubenswrapper[4726]: E1001 16:36:34.464177 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db561a3-0344-4346-8276-cc7770b5efac" containerName="extract-utilities" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.464194 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db561a3-0344-4346-8276-cc7770b5efac" containerName="extract-utilities" Oct 01 16:36:34 crc kubenswrapper[4726]: E1001 16:36:34.464234 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" containerName="extract-content" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.464246 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" containerName="extract-content" Oct 01 16:36:34 crc kubenswrapper[4726]: E1001 16:36:34.464266 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db561a3-0344-4346-8276-cc7770b5efac" containerName="registry-server" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.464280 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db561a3-0344-4346-8276-cc7770b5efac" containerName="registry-server" Oct 01 16:36:34 crc kubenswrapper[4726]: E1001 16:36:34.464313 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" containerName="extract-utilities" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.464324 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" containerName="extract-utilities" Oct 01 16:36:34 crc kubenswrapper[4726]: E1001 16:36:34.464349 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db561a3-0344-4346-8276-cc7770b5efac" containerName="extract-content" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.464359 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db561a3-0344-4346-8276-cc7770b5efac" containerName="extract-content" Oct 01 16:36:34 crc kubenswrapper[4726]: E1001 16:36:34.464375 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" containerName="registry-server" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.464384 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" containerName="registry-server" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.464701 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="7db561a3-0344-4346-8276-cc7770b5efac" containerName="registry-server" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.464744 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea6455f-dafb-46d1-94e4-8ed5ff94ab97" containerName="registry-server" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.466868 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.471413 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-px6g2"] Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.626710 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79acb009-2e26-48ac-9fa0-8d04e35887a4-utilities\") pod \"community-operators-px6g2\" (UID: \"79acb009-2e26-48ac-9fa0-8d04e35887a4\") " pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.626830 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7d68\" (UniqueName: \"kubernetes.io/projected/79acb009-2e26-48ac-9fa0-8d04e35887a4-kube-api-access-w7d68\") pod \"community-operators-px6g2\" (UID: \"79acb009-2e26-48ac-9fa0-8d04e35887a4\") " pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.626944 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79acb009-2e26-48ac-9fa0-8d04e35887a4-catalog-content\") pod \"community-operators-px6g2\" (UID: \"79acb009-2e26-48ac-9fa0-8d04e35887a4\") " pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.728311 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79acb009-2e26-48ac-9fa0-8d04e35887a4-utilities\") pod \"community-operators-px6g2\" (UID: \"79acb009-2e26-48ac-9fa0-8d04e35887a4\") " pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.728715 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7d68\" (UniqueName: \"kubernetes.io/projected/79acb009-2e26-48ac-9fa0-8d04e35887a4-kube-api-access-w7d68\") pod \"community-operators-px6g2\" (UID: \"79acb009-2e26-48ac-9fa0-8d04e35887a4\") " pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.728815 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79acb009-2e26-48ac-9fa0-8d04e35887a4-utilities\") pod \"community-operators-px6g2\" (UID: \"79acb009-2e26-48ac-9fa0-8d04e35887a4\") " pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.728991 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79acb009-2e26-48ac-9fa0-8d04e35887a4-catalog-content\") pod \"community-operators-px6g2\" (UID: \"79acb009-2e26-48ac-9fa0-8d04e35887a4\") " pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.729340 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79acb009-2e26-48ac-9fa0-8d04e35887a4-catalog-content\") pod \"community-operators-px6g2\" (UID: \"79acb009-2e26-48ac-9fa0-8d04e35887a4\") " pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.749395 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7d68\" (UniqueName: \"kubernetes.io/projected/79acb009-2e26-48ac-9fa0-8d04e35887a4-kube-api-access-w7d68\") pod \"community-operators-px6g2\" (UID: \"79acb009-2e26-48ac-9fa0-8d04e35887a4\") " pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:34 crc kubenswrapper[4726]: I1001 16:36:34.796158 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:35 crc kubenswrapper[4726]: I1001 16:36:35.181771 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-px6g2"] Oct 01 16:36:35 crc kubenswrapper[4726]: I1001 16:36:35.966704 4726 generic.go:334] "Generic (PLEG): container finished" podID="79acb009-2e26-48ac-9fa0-8d04e35887a4" containerID="5d98740dc76e9fb528900050972e9a06ee77ac86bf82e370ff0d9fddf2da0a52" exitCode=0 Oct 01 16:36:35 crc kubenswrapper[4726]: I1001 16:36:35.966802 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-px6g2" event={"ID":"79acb009-2e26-48ac-9fa0-8d04e35887a4","Type":"ContainerDied","Data":"5d98740dc76e9fb528900050972e9a06ee77ac86bf82e370ff0d9fddf2da0a52"} Oct 01 16:36:35 crc kubenswrapper[4726]: I1001 16:36:35.967029 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-px6g2" event={"ID":"79acb009-2e26-48ac-9fa0-8d04e35887a4","Type":"ContainerStarted","Data":"f8f073b564e26fea235ccd5afc9ce43af2607a834efd1c6128825313241aaea7"} Oct 01 16:36:35 crc kubenswrapper[4726]: I1001 16:36:35.969454 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:36:37 crc kubenswrapper[4726]: I1001 16:36:37.987346 4726 generic.go:334] "Generic (PLEG): container finished" podID="79acb009-2e26-48ac-9fa0-8d04e35887a4" containerID="f923869295249ed4eb279a4578f8aea69efa8aa6e5eb6563108cc7bc5204322e" exitCode=0 Oct 01 16:36:37 crc kubenswrapper[4726]: I1001 16:36:37.987454 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-px6g2" event={"ID":"79acb009-2e26-48ac-9fa0-8d04e35887a4","Type":"ContainerDied","Data":"f923869295249ed4eb279a4578f8aea69efa8aa6e5eb6563108cc7bc5204322e"} Oct 01 16:36:38 crc kubenswrapper[4726]: I1001 16:36:38.997639 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-px6g2" event={"ID":"79acb009-2e26-48ac-9fa0-8d04e35887a4","Type":"ContainerStarted","Data":"aac86ec8ed8b88732b982900f8bb9a82b0d05ffba8728c1c8c576f39d9d3363d"} Oct 01 16:36:44 crc kubenswrapper[4726]: I1001 16:36:44.796333 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:44 crc kubenswrapper[4726]: I1001 16:36:44.796911 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:44 crc kubenswrapper[4726]: I1001 16:36:44.846063 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:44 crc kubenswrapper[4726]: I1001 16:36:44.865564 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-px6g2" podStartSLOduration=8.401246376 podStartE2EDuration="10.865546296s" podCreationTimestamp="2025-10-01 16:36:34 +0000 UTC" firstStartedPulling="2025-10-01 16:36:35.969197473 +0000 UTC m=+2248.870750050" lastFinishedPulling="2025-10-01 16:36:38.433497353 +0000 UTC m=+2251.335049970" observedRunningTime="2025-10-01 16:36:39.022036528 +0000 UTC m=+2251.923589105" watchObservedRunningTime="2025-10-01 16:36:44.865546296 +0000 UTC m=+2257.767098873" Oct 01 16:36:45 crc kubenswrapper[4726]: I1001 16:36:45.099869 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:45 crc kubenswrapper[4726]: I1001 16:36:45.155691 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-px6g2"] Oct 01 16:36:47 crc kubenswrapper[4726]: I1001 16:36:47.070316 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-px6g2" podUID="79acb009-2e26-48ac-9fa0-8d04e35887a4" containerName="registry-server" containerID="cri-o://aac86ec8ed8b88732b982900f8bb9a82b0d05ffba8728c1c8c576f39d9d3363d" gracePeriod=2 Oct 01 16:36:47 crc kubenswrapper[4726]: I1001 16:36:47.546353 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:47 crc kubenswrapper[4726]: I1001 16:36:47.690335 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79acb009-2e26-48ac-9fa0-8d04e35887a4-utilities\") pod \"79acb009-2e26-48ac-9fa0-8d04e35887a4\" (UID: \"79acb009-2e26-48ac-9fa0-8d04e35887a4\") " Oct 01 16:36:47 crc kubenswrapper[4726]: I1001 16:36:47.690428 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79acb009-2e26-48ac-9fa0-8d04e35887a4-catalog-content\") pod \"79acb009-2e26-48ac-9fa0-8d04e35887a4\" (UID: \"79acb009-2e26-48ac-9fa0-8d04e35887a4\") " Oct 01 16:36:47 crc kubenswrapper[4726]: I1001 16:36:47.690494 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7d68\" (UniqueName: \"kubernetes.io/projected/79acb009-2e26-48ac-9fa0-8d04e35887a4-kube-api-access-w7d68\") pod \"79acb009-2e26-48ac-9fa0-8d04e35887a4\" (UID: \"79acb009-2e26-48ac-9fa0-8d04e35887a4\") " Oct 01 16:36:47 crc kubenswrapper[4726]: I1001 16:36:47.691396 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79acb009-2e26-48ac-9fa0-8d04e35887a4-utilities" (OuterVolumeSpecName: "utilities") pod "79acb009-2e26-48ac-9fa0-8d04e35887a4" (UID: "79acb009-2e26-48ac-9fa0-8d04e35887a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:36:47 crc kubenswrapper[4726]: I1001 16:36:47.711323 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79acb009-2e26-48ac-9fa0-8d04e35887a4-kube-api-access-w7d68" (OuterVolumeSpecName: "kube-api-access-w7d68") pod "79acb009-2e26-48ac-9fa0-8d04e35887a4" (UID: "79acb009-2e26-48ac-9fa0-8d04e35887a4"). InnerVolumeSpecName "kube-api-access-w7d68". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:36:47 crc kubenswrapper[4726]: I1001 16:36:47.762344 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79acb009-2e26-48ac-9fa0-8d04e35887a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79acb009-2e26-48ac-9fa0-8d04e35887a4" (UID: "79acb009-2e26-48ac-9fa0-8d04e35887a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:36:47 crc kubenswrapper[4726]: I1001 16:36:47.793550 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79acb009-2e26-48ac-9fa0-8d04e35887a4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:36:47 crc kubenswrapper[4726]: I1001 16:36:47.793595 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7d68\" (UniqueName: \"kubernetes.io/projected/79acb009-2e26-48ac-9fa0-8d04e35887a4-kube-api-access-w7d68\") on node \"crc\" DevicePath \"\"" Oct 01 16:36:47 crc kubenswrapper[4726]: I1001 16:36:47.793607 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79acb009-2e26-48ac-9fa0-8d04e35887a4-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.084305 4726 generic.go:334] "Generic (PLEG): container finished" podID="79acb009-2e26-48ac-9fa0-8d04e35887a4" containerID="aac86ec8ed8b88732b982900f8bb9a82b0d05ffba8728c1c8c576f39d9d3363d" exitCode=0 Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.084353 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-px6g2" event={"ID":"79acb009-2e26-48ac-9fa0-8d04e35887a4","Type":"ContainerDied","Data":"aac86ec8ed8b88732b982900f8bb9a82b0d05ffba8728c1c8c576f39d9d3363d"} Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.084383 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-px6g2" event={"ID":"79acb009-2e26-48ac-9fa0-8d04e35887a4","Type":"ContainerDied","Data":"f8f073b564e26fea235ccd5afc9ce43af2607a834efd1c6128825313241aaea7"} Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.084418 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-px6g2" Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.084424 4726 scope.go:117] "RemoveContainer" containerID="aac86ec8ed8b88732b982900f8bb9a82b0d05ffba8728c1c8c576f39d9d3363d" Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.103763 4726 scope.go:117] "RemoveContainer" containerID="f923869295249ed4eb279a4578f8aea69efa8aa6e5eb6563108cc7bc5204322e" Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.123545 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-px6g2"] Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.127263 4726 scope.go:117] "RemoveContainer" containerID="5d98740dc76e9fb528900050972e9a06ee77ac86bf82e370ff0d9fddf2da0a52" Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.137641 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-px6g2"] Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.182362 4726 scope.go:117] "RemoveContainer" containerID="aac86ec8ed8b88732b982900f8bb9a82b0d05ffba8728c1c8c576f39d9d3363d" Oct 01 16:36:48 crc kubenswrapper[4726]: E1001 16:36:48.182732 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aac86ec8ed8b88732b982900f8bb9a82b0d05ffba8728c1c8c576f39d9d3363d\": container with ID starting with aac86ec8ed8b88732b982900f8bb9a82b0d05ffba8728c1c8c576f39d9d3363d not found: ID does not exist" containerID="aac86ec8ed8b88732b982900f8bb9a82b0d05ffba8728c1c8c576f39d9d3363d" Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.182761 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aac86ec8ed8b88732b982900f8bb9a82b0d05ffba8728c1c8c576f39d9d3363d"} err="failed to get container status \"aac86ec8ed8b88732b982900f8bb9a82b0d05ffba8728c1c8c576f39d9d3363d\": rpc error: code = NotFound desc = could not find container \"aac86ec8ed8b88732b982900f8bb9a82b0d05ffba8728c1c8c576f39d9d3363d\": container with ID starting with aac86ec8ed8b88732b982900f8bb9a82b0d05ffba8728c1c8c576f39d9d3363d not found: ID does not exist" Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.182781 4726 scope.go:117] "RemoveContainer" containerID="f923869295249ed4eb279a4578f8aea69efa8aa6e5eb6563108cc7bc5204322e" Oct 01 16:36:48 crc kubenswrapper[4726]: E1001 16:36:48.183170 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f923869295249ed4eb279a4578f8aea69efa8aa6e5eb6563108cc7bc5204322e\": container with ID starting with f923869295249ed4eb279a4578f8aea69efa8aa6e5eb6563108cc7bc5204322e not found: ID does not exist" containerID="f923869295249ed4eb279a4578f8aea69efa8aa6e5eb6563108cc7bc5204322e" Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.183192 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f923869295249ed4eb279a4578f8aea69efa8aa6e5eb6563108cc7bc5204322e"} err="failed to get container status \"f923869295249ed4eb279a4578f8aea69efa8aa6e5eb6563108cc7bc5204322e\": rpc error: code = NotFound desc = could not find container \"f923869295249ed4eb279a4578f8aea69efa8aa6e5eb6563108cc7bc5204322e\": container with ID starting with f923869295249ed4eb279a4578f8aea69efa8aa6e5eb6563108cc7bc5204322e not found: ID does not exist" Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.183208 4726 scope.go:117] "RemoveContainer" containerID="5d98740dc76e9fb528900050972e9a06ee77ac86bf82e370ff0d9fddf2da0a52" Oct 01 16:36:48 crc kubenswrapper[4726]: E1001 16:36:48.183464 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d98740dc76e9fb528900050972e9a06ee77ac86bf82e370ff0d9fddf2da0a52\": container with ID starting with 5d98740dc76e9fb528900050972e9a06ee77ac86bf82e370ff0d9fddf2da0a52 not found: ID does not exist" containerID="5d98740dc76e9fb528900050972e9a06ee77ac86bf82e370ff0d9fddf2da0a52" Oct 01 16:36:48 crc kubenswrapper[4726]: I1001 16:36:48.183490 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d98740dc76e9fb528900050972e9a06ee77ac86bf82e370ff0d9fddf2da0a52"} err="failed to get container status \"5d98740dc76e9fb528900050972e9a06ee77ac86bf82e370ff0d9fddf2da0a52\": rpc error: code = NotFound desc = could not find container \"5d98740dc76e9fb528900050972e9a06ee77ac86bf82e370ff0d9fddf2da0a52\": container with ID starting with 5d98740dc76e9fb528900050972e9a06ee77ac86bf82e370ff0d9fddf2da0a52 not found: ID does not exist" Oct 01 16:36:49 crc kubenswrapper[4726]: I1001 16:36:49.818173 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79acb009-2e26-48ac-9fa0-8d04e35887a4" path="/var/lib/kubelet/pods/79acb009-2e26-48ac-9fa0-8d04e35887a4/volumes" Oct 01 16:36:53 crc kubenswrapper[4726]: I1001 16:36:53.413175 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:36:53 crc kubenswrapper[4726]: I1001 16:36:53.413721 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:36:53 crc kubenswrapper[4726]: I1001 16:36:53.413771 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 16:36:53 crc kubenswrapper[4726]: I1001 16:36:53.414566 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:36:53 crc kubenswrapper[4726]: I1001 16:36:53.414630 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" gracePeriod=600 Oct 01 16:36:53 crc kubenswrapper[4726]: E1001 16:36:53.534136 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:36:54 crc kubenswrapper[4726]: I1001 16:36:54.151373 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" exitCode=0 Oct 01 16:36:54 crc kubenswrapper[4726]: I1001 16:36:54.151436 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c"} Oct 01 16:36:54 crc kubenswrapper[4726]: I1001 16:36:54.151491 4726 scope.go:117] "RemoveContainer" containerID="8ea6a11e2a728a74e08b777157550e1c40f0e83c9ab337602b7696ba954f563a" Oct 01 16:36:54 crc kubenswrapper[4726]: I1001 16:36:54.152317 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:36:54 crc kubenswrapper[4726]: E1001 16:36:54.152887 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:37:05 crc kubenswrapper[4726]: I1001 16:37:05.811262 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:37:05 crc kubenswrapper[4726]: E1001 16:37:05.812117 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:37:18 crc kubenswrapper[4726]: I1001 16:37:18.808190 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:37:18 crc kubenswrapper[4726]: E1001 16:37:18.808947 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:37:30 crc kubenswrapper[4726]: I1001 16:37:30.807950 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:37:30 crc kubenswrapper[4726]: E1001 16:37:30.808771 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:37:45 crc kubenswrapper[4726]: I1001 16:37:45.808358 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:37:45 crc kubenswrapper[4726]: E1001 16:37:45.809405 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:37:57 crc kubenswrapper[4726]: I1001 16:37:57.819260 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:37:57 crc kubenswrapper[4726]: E1001 16:37:57.821367 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:38:12 crc kubenswrapper[4726]: I1001 16:38:12.808648 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:38:12 crc kubenswrapper[4726]: E1001 16:38:12.809909 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:38:27 crc kubenswrapper[4726]: I1001 16:38:27.814604 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:38:27 crc kubenswrapper[4726]: E1001 16:38:27.815377 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:38:35 crc kubenswrapper[4726]: I1001 16:38:35.168686 4726 generic.go:334] "Generic (PLEG): container finished" podID="a66fccbb-d629-4290-a469-066027643cfa" containerID="69980698b0be0d842211ac5ab670385b30af0f68306e71515e421952101a311c" exitCode=0 Oct 01 16:38:35 crc kubenswrapper[4726]: I1001 16:38:35.168769 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" event={"ID":"a66fccbb-d629-4290-a469-066027643cfa","Type":"ContainerDied","Data":"69980698b0be0d842211ac5ab670385b30af0f68306e71515e421952101a311c"} Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.571070 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.688889 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-libvirt-secret-0\") pod \"a66fccbb-d629-4290-a469-066027643cfa\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.688997 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dm8lw\" (UniqueName: \"kubernetes.io/projected/a66fccbb-d629-4290-a469-066027643cfa-kube-api-access-dm8lw\") pod \"a66fccbb-d629-4290-a469-066027643cfa\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.689042 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-inventory\") pod \"a66fccbb-d629-4290-a469-066027643cfa\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.689078 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-libvirt-combined-ca-bundle\") pod \"a66fccbb-d629-4290-a469-066027643cfa\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.689125 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-ssh-key\") pod \"a66fccbb-d629-4290-a469-066027643cfa\" (UID: \"a66fccbb-d629-4290-a469-066027643cfa\") " Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.694290 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "a66fccbb-d629-4290-a469-066027643cfa" (UID: "a66fccbb-d629-4290-a469-066027643cfa"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.696387 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a66fccbb-d629-4290-a469-066027643cfa-kube-api-access-dm8lw" (OuterVolumeSpecName: "kube-api-access-dm8lw") pod "a66fccbb-d629-4290-a469-066027643cfa" (UID: "a66fccbb-d629-4290-a469-066027643cfa"). InnerVolumeSpecName "kube-api-access-dm8lw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.716112 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-inventory" (OuterVolumeSpecName: "inventory") pod "a66fccbb-d629-4290-a469-066027643cfa" (UID: "a66fccbb-d629-4290-a469-066027643cfa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.716988 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a66fccbb-d629-4290-a469-066027643cfa" (UID: "a66fccbb-d629-4290-a469-066027643cfa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.720392 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "a66fccbb-d629-4290-a469-066027643cfa" (UID: "a66fccbb-d629-4290-a469-066027643cfa"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.791742 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dm8lw\" (UniqueName: \"kubernetes.io/projected/a66fccbb-d629-4290-a469-066027643cfa-kube-api-access-dm8lw\") on node \"crc\" DevicePath \"\"" Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.791789 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.791801 4726 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.791812 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:38:36 crc kubenswrapper[4726]: I1001 16:38:36.791844 4726 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a66fccbb-d629-4290-a469-066027643cfa-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.190916 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" event={"ID":"a66fccbb-d629-4290-a469-066027643cfa","Type":"ContainerDied","Data":"e0bbe0fe0d459b52c1bfda2a7a7d600699bb7d3e48c52928b4d50909bfcb5b15"} Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.190966 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0bbe0fe0d459b52c1bfda2a7a7d600699bb7d3e48c52928b4d50909bfcb5b15" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.191002 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-svtfg" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.363540 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq"] Oct 01 16:38:37 crc kubenswrapper[4726]: E1001 16:38:37.363936 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66fccbb-d629-4290-a469-066027643cfa" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.363954 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66fccbb-d629-4290-a469-066027643cfa" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 01 16:38:37 crc kubenswrapper[4726]: E1001 16:38:37.363968 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79acb009-2e26-48ac-9fa0-8d04e35887a4" containerName="extract-content" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.363975 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="79acb009-2e26-48ac-9fa0-8d04e35887a4" containerName="extract-content" Oct 01 16:38:37 crc kubenswrapper[4726]: E1001 16:38:37.363994 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79acb009-2e26-48ac-9fa0-8d04e35887a4" containerName="extract-utilities" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.364000 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="79acb009-2e26-48ac-9fa0-8d04e35887a4" containerName="extract-utilities" Oct 01 16:38:37 crc kubenswrapper[4726]: E1001 16:38:37.364011 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79acb009-2e26-48ac-9fa0-8d04e35887a4" containerName="registry-server" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.364017 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="79acb009-2e26-48ac-9fa0-8d04e35887a4" containerName="registry-server" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.364201 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="79acb009-2e26-48ac-9fa0-8d04e35887a4" containerName="registry-server" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.364230 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66fccbb-d629-4290-a469-066027643cfa" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.364825 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.371387 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.371745 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.371745 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.371765 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.371776 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.375889 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.376211 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.393573 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq"] Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.404481 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.404530 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.404566 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.404693 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.404720 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpvhh\" (UniqueName: \"kubernetes.io/projected/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-kube-api-access-cpvhh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.404740 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.404854 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.404890 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.404918 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.506111 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.506148 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpvhh\" (UniqueName: \"kubernetes.io/projected/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-kube-api-access-cpvhh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.506173 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.506242 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.506263 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.506282 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.506323 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.506499 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.506560 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.506908 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.510415 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.510890 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.512209 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.512557 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.515009 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.516211 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.518906 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.528186 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpvhh\" (UniqueName: \"kubernetes.io/projected/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-kube-api-access-cpvhh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-d77vq\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:37 crc kubenswrapper[4726]: I1001 16:38:37.696345 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:38:38 crc kubenswrapper[4726]: I1001 16:38:38.244780 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq"] Oct 01 16:38:38 crc kubenswrapper[4726]: I1001 16:38:38.808339 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:38:38 crc kubenswrapper[4726]: E1001 16:38:38.809046 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:38:39 crc kubenswrapper[4726]: I1001 16:38:39.209615 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" event={"ID":"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d","Type":"ContainerStarted","Data":"06b9f0cab3d20012db8877b29132eec5ad5da47123b77471936d8bb1c3960b9b"} Oct 01 16:38:39 crc kubenswrapper[4726]: I1001 16:38:39.209861 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" event={"ID":"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d","Type":"ContainerStarted","Data":"9cf50e0c33fc0816e63330a0ae162edeb98d63ed3f3e88d180d54115509b7462"} Oct 01 16:38:39 crc kubenswrapper[4726]: I1001 16:38:39.232245 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" podStartSLOduration=1.742942064 podStartE2EDuration="2.23222607s" podCreationTimestamp="2025-10-01 16:38:37 +0000 UTC" firstStartedPulling="2025-10-01 16:38:38.253634769 +0000 UTC m=+2371.155187346" lastFinishedPulling="2025-10-01 16:38:38.742918755 +0000 UTC m=+2371.644471352" observedRunningTime="2025-10-01 16:38:39.226698269 +0000 UTC m=+2372.128250846" watchObservedRunningTime="2025-10-01 16:38:39.23222607 +0000 UTC m=+2372.133778647" Oct 01 16:38:49 crc kubenswrapper[4726]: I1001 16:38:49.809005 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:38:49 crc kubenswrapper[4726]: E1001 16:38:49.809821 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:39:00 crc kubenswrapper[4726]: I1001 16:39:00.809548 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:39:00 crc kubenswrapper[4726]: E1001 16:39:00.810212 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:39:12 crc kubenswrapper[4726]: I1001 16:39:12.809205 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:39:12 crc kubenswrapper[4726]: E1001 16:39:12.810314 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:39:26 crc kubenswrapper[4726]: I1001 16:39:26.808695 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:39:26 crc kubenswrapper[4726]: E1001 16:39:26.809485 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:39:39 crc kubenswrapper[4726]: I1001 16:39:39.808700 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:39:39 crc kubenswrapper[4726]: E1001 16:39:39.809995 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:39:51 crc kubenswrapper[4726]: I1001 16:39:51.807997 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:39:51 crc kubenswrapper[4726]: E1001 16:39:51.812220 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:40:04 crc kubenswrapper[4726]: I1001 16:40:04.200692 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j69v4"] Oct 01 16:40:04 crc kubenswrapper[4726]: I1001 16:40:04.203268 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:04 crc kubenswrapper[4726]: I1001 16:40:04.222586 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j69v4"] Oct 01 16:40:04 crc kubenswrapper[4726]: I1001 16:40:04.294004 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0203c4a8-ad55-40de-9b16-8129a74af6f6-catalog-content\") pod \"redhat-operators-j69v4\" (UID: \"0203c4a8-ad55-40de-9b16-8129a74af6f6\") " pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:04 crc kubenswrapper[4726]: I1001 16:40:04.294297 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz6fs\" (UniqueName: \"kubernetes.io/projected/0203c4a8-ad55-40de-9b16-8129a74af6f6-kube-api-access-xz6fs\") pod \"redhat-operators-j69v4\" (UID: \"0203c4a8-ad55-40de-9b16-8129a74af6f6\") " pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:04 crc kubenswrapper[4726]: I1001 16:40:04.294418 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0203c4a8-ad55-40de-9b16-8129a74af6f6-utilities\") pod \"redhat-operators-j69v4\" (UID: \"0203c4a8-ad55-40de-9b16-8129a74af6f6\") " pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:04 crc kubenswrapper[4726]: I1001 16:40:04.396387 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0203c4a8-ad55-40de-9b16-8129a74af6f6-utilities\") pod \"redhat-operators-j69v4\" (UID: \"0203c4a8-ad55-40de-9b16-8129a74af6f6\") " pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:04 crc kubenswrapper[4726]: I1001 16:40:04.396787 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0203c4a8-ad55-40de-9b16-8129a74af6f6-catalog-content\") pod \"redhat-operators-j69v4\" (UID: \"0203c4a8-ad55-40de-9b16-8129a74af6f6\") " pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:04 crc kubenswrapper[4726]: I1001 16:40:04.396808 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0203c4a8-ad55-40de-9b16-8129a74af6f6-utilities\") pod \"redhat-operators-j69v4\" (UID: \"0203c4a8-ad55-40de-9b16-8129a74af6f6\") " pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:04 crc kubenswrapper[4726]: I1001 16:40:04.397033 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0203c4a8-ad55-40de-9b16-8129a74af6f6-catalog-content\") pod \"redhat-operators-j69v4\" (UID: \"0203c4a8-ad55-40de-9b16-8129a74af6f6\") " pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:04 crc kubenswrapper[4726]: I1001 16:40:04.397304 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz6fs\" (UniqueName: \"kubernetes.io/projected/0203c4a8-ad55-40de-9b16-8129a74af6f6-kube-api-access-xz6fs\") pod \"redhat-operators-j69v4\" (UID: \"0203c4a8-ad55-40de-9b16-8129a74af6f6\") " pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:04 crc kubenswrapper[4726]: I1001 16:40:04.419618 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz6fs\" (UniqueName: \"kubernetes.io/projected/0203c4a8-ad55-40de-9b16-8129a74af6f6-kube-api-access-xz6fs\") pod \"redhat-operators-j69v4\" (UID: \"0203c4a8-ad55-40de-9b16-8129a74af6f6\") " pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:04 crc kubenswrapper[4726]: I1001 16:40:04.524155 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:05 crc kubenswrapper[4726]: I1001 16:40:04.974710 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j69v4"] Oct 01 16:40:05 crc kubenswrapper[4726]: W1001 16:40:04.977359 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0203c4a8_ad55_40de_9b16_8129a74af6f6.slice/crio-8c57cf9d56855394fcaa6c9ee656916f7a3606a78c731dc8ac4843120c8880d5 WatchSource:0}: Error finding container 8c57cf9d56855394fcaa6c9ee656916f7a3606a78c731dc8ac4843120c8880d5: Status 404 returned error can't find the container with id 8c57cf9d56855394fcaa6c9ee656916f7a3606a78c731dc8ac4843120c8880d5 Oct 01 16:40:05 crc kubenswrapper[4726]: I1001 16:40:05.208227 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69v4" event={"ID":"0203c4a8-ad55-40de-9b16-8129a74af6f6","Type":"ContainerStarted","Data":"8c57cf9d56855394fcaa6c9ee656916f7a3606a78c731dc8ac4843120c8880d5"} Oct 01 16:40:05 crc kubenswrapper[4726]: I1001 16:40:05.808618 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:40:05 crc kubenswrapper[4726]: E1001 16:40:05.809093 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:40:06 crc kubenswrapper[4726]: I1001 16:40:06.221019 4726 generic.go:334] "Generic (PLEG): container finished" podID="0203c4a8-ad55-40de-9b16-8129a74af6f6" containerID="f334b2a4e6ed9b30ae343e3f468136b63f31d0e58b2db62972b2e4f08f27f6d0" exitCode=0 Oct 01 16:40:06 crc kubenswrapper[4726]: I1001 16:40:06.221131 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69v4" event={"ID":"0203c4a8-ad55-40de-9b16-8129a74af6f6","Type":"ContainerDied","Data":"f334b2a4e6ed9b30ae343e3f468136b63f31d0e58b2db62972b2e4f08f27f6d0"} Oct 01 16:40:07 crc kubenswrapper[4726]: I1001 16:40:07.231979 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69v4" event={"ID":"0203c4a8-ad55-40de-9b16-8129a74af6f6","Type":"ContainerStarted","Data":"c4e7187d286a3a2f191b18a3d49932a59723b659f855105e9c2b2ef86b368f0b"} Oct 01 16:40:11 crc kubenswrapper[4726]: I1001 16:40:11.271443 4726 generic.go:334] "Generic (PLEG): container finished" podID="0203c4a8-ad55-40de-9b16-8129a74af6f6" containerID="c4e7187d286a3a2f191b18a3d49932a59723b659f855105e9c2b2ef86b368f0b" exitCode=0 Oct 01 16:40:11 crc kubenswrapper[4726]: I1001 16:40:11.271524 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69v4" event={"ID":"0203c4a8-ad55-40de-9b16-8129a74af6f6","Type":"ContainerDied","Data":"c4e7187d286a3a2f191b18a3d49932a59723b659f855105e9c2b2ef86b368f0b"} Oct 01 16:40:12 crc kubenswrapper[4726]: I1001 16:40:12.285689 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69v4" event={"ID":"0203c4a8-ad55-40de-9b16-8129a74af6f6","Type":"ContainerStarted","Data":"554e67f732c2336c2a2d45328f6c9eb594fee29e3d10c7409cc86108b1d7c03b"} Oct 01 16:40:12 crc kubenswrapper[4726]: I1001 16:40:12.309280 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j69v4" podStartSLOduration=2.545156737 podStartE2EDuration="8.309259795s" podCreationTimestamp="2025-10-01 16:40:04 +0000 UTC" firstStartedPulling="2025-10-01 16:40:06.223919544 +0000 UTC m=+2459.125472161" lastFinishedPulling="2025-10-01 16:40:11.988022642 +0000 UTC m=+2464.889575219" observedRunningTime="2025-10-01 16:40:12.307343379 +0000 UTC m=+2465.208895956" watchObservedRunningTime="2025-10-01 16:40:12.309259795 +0000 UTC m=+2465.210812372" Oct 01 16:40:14 crc kubenswrapper[4726]: I1001 16:40:14.524777 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:14 crc kubenswrapper[4726]: I1001 16:40:14.525417 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:15 crc kubenswrapper[4726]: I1001 16:40:15.580696 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j69v4" podUID="0203c4a8-ad55-40de-9b16-8129a74af6f6" containerName="registry-server" probeResult="failure" output=< Oct 01 16:40:15 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Oct 01 16:40:15 crc kubenswrapper[4726]: > Oct 01 16:40:16 crc kubenswrapper[4726]: I1001 16:40:16.807997 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:40:16 crc kubenswrapper[4726]: E1001 16:40:16.808511 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:40:24 crc kubenswrapper[4726]: I1001 16:40:24.596752 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:24 crc kubenswrapper[4726]: I1001 16:40:24.661264 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:24 crc kubenswrapper[4726]: I1001 16:40:24.826916 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j69v4"] Oct 01 16:40:26 crc kubenswrapper[4726]: I1001 16:40:26.437133 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j69v4" podUID="0203c4a8-ad55-40de-9b16-8129a74af6f6" containerName="registry-server" containerID="cri-o://554e67f732c2336c2a2d45328f6c9eb594fee29e3d10c7409cc86108b1d7c03b" gracePeriod=2 Oct 01 16:40:26 crc kubenswrapper[4726]: I1001 16:40:26.846361 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:26 crc kubenswrapper[4726]: I1001 16:40:26.975856 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz6fs\" (UniqueName: \"kubernetes.io/projected/0203c4a8-ad55-40de-9b16-8129a74af6f6-kube-api-access-xz6fs\") pod \"0203c4a8-ad55-40de-9b16-8129a74af6f6\" (UID: \"0203c4a8-ad55-40de-9b16-8129a74af6f6\") " Oct 01 16:40:26 crc kubenswrapper[4726]: I1001 16:40:26.975970 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0203c4a8-ad55-40de-9b16-8129a74af6f6-utilities\") pod \"0203c4a8-ad55-40de-9b16-8129a74af6f6\" (UID: \"0203c4a8-ad55-40de-9b16-8129a74af6f6\") " Oct 01 16:40:26 crc kubenswrapper[4726]: I1001 16:40:26.976042 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0203c4a8-ad55-40de-9b16-8129a74af6f6-catalog-content\") pod \"0203c4a8-ad55-40de-9b16-8129a74af6f6\" (UID: \"0203c4a8-ad55-40de-9b16-8129a74af6f6\") " Oct 01 16:40:26 crc kubenswrapper[4726]: I1001 16:40:26.979418 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0203c4a8-ad55-40de-9b16-8129a74af6f6-utilities" (OuterVolumeSpecName: "utilities") pod "0203c4a8-ad55-40de-9b16-8129a74af6f6" (UID: "0203c4a8-ad55-40de-9b16-8129a74af6f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:40:26 crc kubenswrapper[4726]: I1001 16:40:26.982083 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0203c4a8-ad55-40de-9b16-8129a74af6f6-kube-api-access-xz6fs" (OuterVolumeSpecName: "kube-api-access-xz6fs") pod "0203c4a8-ad55-40de-9b16-8129a74af6f6" (UID: "0203c4a8-ad55-40de-9b16-8129a74af6f6"). InnerVolumeSpecName "kube-api-access-xz6fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.096991 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz6fs\" (UniqueName: \"kubernetes.io/projected/0203c4a8-ad55-40de-9b16-8129a74af6f6-kube-api-access-xz6fs\") on node \"crc\" DevicePath \"\"" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.097039 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0203c4a8-ad55-40de-9b16-8129a74af6f6-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.100360 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0203c4a8-ad55-40de-9b16-8129a74af6f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0203c4a8-ad55-40de-9b16-8129a74af6f6" (UID: "0203c4a8-ad55-40de-9b16-8129a74af6f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.200100 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0203c4a8-ad55-40de-9b16-8129a74af6f6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.450697 4726 generic.go:334] "Generic (PLEG): container finished" podID="0203c4a8-ad55-40de-9b16-8129a74af6f6" containerID="554e67f732c2336c2a2d45328f6c9eb594fee29e3d10c7409cc86108b1d7c03b" exitCode=0 Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.450742 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69v4" event={"ID":"0203c4a8-ad55-40de-9b16-8129a74af6f6","Type":"ContainerDied","Data":"554e67f732c2336c2a2d45328f6c9eb594fee29e3d10c7409cc86108b1d7c03b"} Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.450768 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69v4" event={"ID":"0203c4a8-ad55-40de-9b16-8129a74af6f6","Type":"ContainerDied","Data":"8c57cf9d56855394fcaa6c9ee656916f7a3606a78c731dc8ac4843120c8880d5"} Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.450785 4726 scope.go:117] "RemoveContainer" containerID="554e67f732c2336c2a2d45328f6c9eb594fee29e3d10c7409cc86108b1d7c03b" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.450920 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j69v4" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.488659 4726 scope.go:117] "RemoveContainer" containerID="c4e7187d286a3a2f191b18a3d49932a59723b659f855105e9c2b2ef86b368f0b" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.509481 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j69v4"] Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.518392 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j69v4"] Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.539373 4726 scope.go:117] "RemoveContainer" containerID="f334b2a4e6ed9b30ae343e3f468136b63f31d0e58b2db62972b2e4f08f27f6d0" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.569768 4726 scope.go:117] "RemoveContainer" containerID="554e67f732c2336c2a2d45328f6c9eb594fee29e3d10c7409cc86108b1d7c03b" Oct 01 16:40:27 crc kubenswrapper[4726]: E1001 16:40:27.570376 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"554e67f732c2336c2a2d45328f6c9eb594fee29e3d10c7409cc86108b1d7c03b\": container with ID starting with 554e67f732c2336c2a2d45328f6c9eb594fee29e3d10c7409cc86108b1d7c03b not found: ID does not exist" containerID="554e67f732c2336c2a2d45328f6c9eb594fee29e3d10c7409cc86108b1d7c03b" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.570414 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"554e67f732c2336c2a2d45328f6c9eb594fee29e3d10c7409cc86108b1d7c03b"} err="failed to get container status \"554e67f732c2336c2a2d45328f6c9eb594fee29e3d10c7409cc86108b1d7c03b\": rpc error: code = NotFound desc = could not find container \"554e67f732c2336c2a2d45328f6c9eb594fee29e3d10c7409cc86108b1d7c03b\": container with ID starting with 554e67f732c2336c2a2d45328f6c9eb594fee29e3d10c7409cc86108b1d7c03b not found: ID does not exist" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.570439 4726 scope.go:117] "RemoveContainer" containerID="c4e7187d286a3a2f191b18a3d49932a59723b659f855105e9c2b2ef86b368f0b" Oct 01 16:40:27 crc kubenswrapper[4726]: E1001 16:40:27.571020 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4e7187d286a3a2f191b18a3d49932a59723b659f855105e9c2b2ef86b368f0b\": container with ID starting with c4e7187d286a3a2f191b18a3d49932a59723b659f855105e9c2b2ef86b368f0b not found: ID does not exist" containerID="c4e7187d286a3a2f191b18a3d49932a59723b659f855105e9c2b2ef86b368f0b" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.571075 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4e7187d286a3a2f191b18a3d49932a59723b659f855105e9c2b2ef86b368f0b"} err="failed to get container status \"c4e7187d286a3a2f191b18a3d49932a59723b659f855105e9c2b2ef86b368f0b\": rpc error: code = NotFound desc = could not find container \"c4e7187d286a3a2f191b18a3d49932a59723b659f855105e9c2b2ef86b368f0b\": container with ID starting with c4e7187d286a3a2f191b18a3d49932a59723b659f855105e9c2b2ef86b368f0b not found: ID does not exist" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.571103 4726 scope.go:117] "RemoveContainer" containerID="f334b2a4e6ed9b30ae343e3f468136b63f31d0e58b2db62972b2e4f08f27f6d0" Oct 01 16:40:27 crc kubenswrapper[4726]: E1001 16:40:27.571429 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f334b2a4e6ed9b30ae343e3f468136b63f31d0e58b2db62972b2e4f08f27f6d0\": container with ID starting with f334b2a4e6ed9b30ae343e3f468136b63f31d0e58b2db62972b2e4f08f27f6d0 not found: ID does not exist" containerID="f334b2a4e6ed9b30ae343e3f468136b63f31d0e58b2db62972b2e4f08f27f6d0" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.571463 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f334b2a4e6ed9b30ae343e3f468136b63f31d0e58b2db62972b2e4f08f27f6d0"} err="failed to get container status \"f334b2a4e6ed9b30ae343e3f468136b63f31d0e58b2db62972b2e4f08f27f6d0\": rpc error: code = NotFound desc = could not find container \"f334b2a4e6ed9b30ae343e3f468136b63f31d0e58b2db62972b2e4f08f27f6d0\": container with ID starting with f334b2a4e6ed9b30ae343e3f468136b63f31d0e58b2db62972b2e4f08f27f6d0 not found: ID does not exist" Oct 01 16:40:27 crc kubenswrapper[4726]: I1001 16:40:27.825153 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0203c4a8-ad55-40de-9b16-8129a74af6f6" path="/var/lib/kubelet/pods/0203c4a8-ad55-40de-9b16-8129a74af6f6/volumes" Oct 01 16:40:31 crc kubenswrapper[4726]: I1001 16:40:31.807942 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:40:31 crc kubenswrapper[4726]: E1001 16:40:31.808651 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:40:42 crc kubenswrapper[4726]: I1001 16:40:42.808596 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:40:42 crc kubenswrapper[4726]: E1001 16:40:42.809339 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:40:56 crc kubenswrapper[4726]: I1001 16:40:56.809580 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:40:56 crc kubenswrapper[4726]: E1001 16:40:56.811569 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:41:08 crc kubenswrapper[4726]: I1001 16:41:08.809235 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:41:08 crc kubenswrapper[4726]: E1001 16:41:08.810099 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:41:23 crc kubenswrapper[4726]: I1001 16:41:23.808318 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:41:23 crc kubenswrapper[4726]: E1001 16:41:23.809019 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:41:38 crc kubenswrapper[4726]: I1001 16:41:38.808124 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:41:38 crc kubenswrapper[4726]: E1001 16:41:38.808984 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:41:49 crc kubenswrapper[4726]: I1001 16:41:49.260238 4726 generic.go:334] "Generic (PLEG): container finished" podID="d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d" containerID="06b9f0cab3d20012db8877b29132eec5ad5da47123b77471936d8bb1c3960b9b" exitCode=0 Oct 01 16:41:49 crc kubenswrapper[4726]: I1001 16:41:49.260299 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" event={"ID":"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d","Type":"ContainerDied","Data":"06b9f0cab3d20012db8877b29132eec5ad5da47123b77471936d8bb1c3960b9b"} Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.776481 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.848656 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-cell1-compute-config-1\") pod \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.848710 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-extra-config-0\") pod \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.848749 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-ssh-key\") pod \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.873955 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d" (UID: "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.881828 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d" (UID: "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.883861 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d" (UID: "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.950921 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-migration-ssh-key-1\") pod \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.951249 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-inventory\") pod \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.951525 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-migration-ssh-key-0\") pod \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.951646 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-cell1-compute-config-0\") pod \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.951780 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpvhh\" (UniqueName: \"kubernetes.io/projected/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-kube-api-access-cpvhh\") pod \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.951982 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-combined-ca-bundle\") pod \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\" (UID: \"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d\") " Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.952737 4726 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.952845 4726 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.952979 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.956671 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-kube-api-access-cpvhh" (OuterVolumeSpecName: "kube-api-access-cpvhh") pod "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d" (UID: "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d"). InnerVolumeSpecName "kube-api-access-cpvhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.958917 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d" (UID: "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.984518 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d" (UID: "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.991625 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-inventory" (OuterVolumeSpecName: "inventory") pod "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d" (UID: "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.992476 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d" (UID: "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:41:50 crc kubenswrapper[4726]: I1001 16:41:50.997737 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d" (UID: "d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.054874 4726 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.054901 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpvhh\" (UniqueName: \"kubernetes.io/projected/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-kube-api-access-cpvhh\") on node \"crc\" DevicePath \"\"" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.054910 4726 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.054940 4726 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.054949 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.054957 4726 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.284783 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" event={"ID":"d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d","Type":"ContainerDied","Data":"9cf50e0c33fc0816e63330a0ae162edeb98d63ed3f3e88d180d54115509b7462"} Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.285299 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cf50e0c33fc0816e63330a0ae162edeb98d63ed3f3e88d180d54115509b7462" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.284879 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-d77vq" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.421946 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k"] Oct 01 16:41:51 crc kubenswrapper[4726]: E1001 16:41:51.422327 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.422344 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 01 16:41:51 crc kubenswrapper[4726]: E1001 16:41:51.422363 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0203c4a8-ad55-40de-9b16-8129a74af6f6" containerName="registry-server" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.422369 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0203c4a8-ad55-40de-9b16-8129a74af6f6" containerName="registry-server" Oct 01 16:41:51 crc kubenswrapper[4726]: E1001 16:41:51.422390 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0203c4a8-ad55-40de-9b16-8129a74af6f6" containerName="extract-utilities" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.422397 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0203c4a8-ad55-40de-9b16-8129a74af6f6" containerName="extract-utilities" Oct 01 16:41:51 crc kubenswrapper[4726]: E1001 16:41:51.422420 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0203c4a8-ad55-40de-9b16-8129a74af6f6" containerName="extract-content" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.422430 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0203c4a8-ad55-40de-9b16-8129a74af6f6" containerName="extract-content" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.422642 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0203c4a8-ad55-40de-9b16-8129a74af6f6" containerName="registry-server" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.422663 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.423409 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.428406 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zq8rh" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.428755 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.431324 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.436247 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.438289 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.440803 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k"] Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.566918 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.567134 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.567183 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.567242 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.567285 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42n9h\" (UniqueName: \"kubernetes.io/projected/376afdd4-d312-464e-9627-42656734a785-kube-api-access-42n9h\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.567560 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.567664 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.668960 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.669032 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.669820 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.669858 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42n9h\" (UniqueName: \"kubernetes.io/projected/376afdd4-d312-464e-9627-42656734a785-kube-api-access-42n9h\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.669908 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.669934 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.670034 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.674497 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.674546 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.674814 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.674994 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.675039 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.675535 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.686336 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42n9h\" (UniqueName: \"kubernetes.io/projected/376afdd4-d312-464e-9627-42656734a785-kube-api-access-42n9h\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:51 crc kubenswrapper[4726]: I1001 16:41:51.760583 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:41:52 crc kubenswrapper[4726]: I1001 16:41:52.323845 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k"] Oct 01 16:41:52 crc kubenswrapper[4726]: I1001 16:41:52.331184 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:41:53 crc kubenswrapper[4726]: I1001 16:41:53.304451 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" event={"ID":"376afdd4-d312-464e-9627-42656734a785","Type":"ContainerStarted","Data":"e5318c77ba0a642d17646c4363d053441c272b532132a8f1bde14efcee09544d"} Oct 01 16:41:53 crc kubenswrapper[4726]: I1001 16:41:53.304963 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" event={"ID":"376afdd4-d312-464e-9627-42656734a785","Type":"ContainerStarted","Data":"fa03650fbcf1872260e58d23fed9dc17fb0db713a7d1a2d4de80a9c744f53ad1"} Oct 01 16:41:53 crc kubenswrapper[4726]: I1001 16:41:53.322350 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" podStartSLOduration=1.8121364519999998 podStartE2EDuration="2.322331957s" podCreationTimestamp="2025-10-01 16:41:51 +0000 UTC" firstStartedPulling="2025-10-01 16:41:52.330922381 +0000 UTC m=+2565.232474958" lastFinishedPulling="2025-10-01 16:41:52.841117886 +0000 UTC m=+2565.742670463" observedRunningTime="2025-10-01 16:41:53.320305858 +0000 UTC m=+2566.221858445" watchObservedRunningTime="2025-10-01 16:41:53.322331957 +0000 UTC m=+2566.223884534" Oct 01 16:41:53 crc kubenswrapper[4726]: I1001 16:41:53.809401 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:41:54 crc kubenswrapper[4726]: I1001 16:41:54.319512 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"58493af1875a9a504d29f76fe2259664f298bbadd22e008df04c6a644f787e87"} Oct 01 16:44:17 crc kubenswrapper[4726]: I1001 16:44:17.798194 4726 generic.go:334] "Generic (PLEG): container finished" podID="376afdd4-d312-464e-9627-42656734a785" containerID="e5318c77ba0a642d17646c4363d053441c272b532132a8f1bde14efcee09544d" exitCode=0 Oct 01 16:44:17 crc kubenswrapper[4726]: I1001 16:44:17.798656 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" event={"ID":"376afdd4-d312-464e-9627-42656734a785","Type":"ContainerDied","Data":"e5318c77ba0a642d17646c4363d053441c272b532132a8f1bde14efcee09544d"} Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.232211 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.326392 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42n9h\" (UniqueName: \"kubernetes.io/projected/376afdd4-d312-464e-9627-42656734a785-kube-api-access-42n9h\") pod \"376afdd4-d312-464e-9627-42656734a785\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.326486 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-1\") pod \"376afdd4-d312-464e-9627-42656734a785\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.326589 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-telemetry-combined-ca-bundle\") pod \"376afdd4-d312-464e-9627-42656734a785\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.326609 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-0\") pod \"376afdd4-d312-464e-9627-42656734a785\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.326635 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-inventory\") pod \"376afdd4-d312-464e-9627-42656734a785\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.326674 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ssh-key\") pod \"376afdd4-d312-464e-9627-42656734a785\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.326704 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-2\") pod \"376afdd4-d312-464e-9627-42656734a785\" (UID: \"376afdd4-d312-464e-9627-42656734a785\") " Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.333186 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "376afdd4-d312-464e-9627-42656734a785" (UID: "376afdd4-d312-464e-9627-42656734a785"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.333683 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/376afdd4-d312-464e-9627-42656734a785-kube-api-access-42n9h" (OuterVolumeSpecName: "kube-api-access-42n9h") pod "376afdd4-d312-464e-9627-42656734a785" (UID: "376afdd4-d312-464e-9627-42656734a785"). InnerVolumeSpecName "kube-api-access-42n9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.356361 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "376afdd4-d312-464e-9627-42656734a785" (UID: "376afdd4-d312-464e-9627-42656734a785"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.357755 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-inventory" (OuterVolumeSpecName: "inventory") pod "376afdd4-d312-464e-9627-42656734a785" (UID: "376afdd4-d312-464e-9627-42656734a785"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.358256 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "376afdd4-d312-464e-9627-42656734a785" (UID: "376afdd4-d312-464e-9627-42656734a785"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.359079 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "376afdd4-d312-464e-9627-42656734a785" (UID: "376afdd4-d312-464e-9627-42656734a785"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.366175 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "376afdd4-d312-464e-9627-42656734a785" (UID: "376afdd4-d312-464e-9627-42656734a785"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.429230 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42n9h\" (UniqueName: \"kubernetes.io/projected/376afdd4-d312-464e-9627-42656734a785-kube-api-access-42n9h\") on node \"crc\" DevicePath \"\"" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.429263 4726 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.429274 4726 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.429284 4726 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.429293 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.429302 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.429310 4726 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/376afdd4-d312-464e-9627-42656734a785-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.842745 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" event={"ID":"376afdd4-d312-464e-9627-42656734a785","Type":"ContainerDied","Data":"fa03650fbcf1872260e58d23fed9dc17fb0db713a7d1a2d4de80a9c744f53ad1"} Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.842795 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa03650fbcf1872260e58d23fed9dc17fb0db713a7d1a2d4de80a9c744f53ad1" Oct 01 16:44:19 crc kubenswrapper[4726]: I1001 16:44:19.842762 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k" Oct 01 16:44:23 crc kubenswrapper[4726]: I1001 16:44:23.413837 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:44:23 crc kubenswrapper[4726]: I1001 16:44:23.414553 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:44:53 crc kubenswrapper[4726]: I1001 16:44:53.413344 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:44:53 crc kubenswrapper[4726]: I1001 16:44:53.414096 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.164085 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868"] Oct 01 16:45:00 crc kubenswrapper[4726]: E1001 16:45:00.165584 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376afdd4-d312-464e-9627-42656734a785" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.165617 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="376afdd4-d312-464e-9627-42656734a785" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.166124 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="376afdd4-d312-464e-9627-42656734a785" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.167501 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.169625 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.169920 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.194210 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868"] Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.296721 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhqb6\" (UniqueName: \"kubernetes.io/projected/72d4a425-8329-455c-8075-2c1657fec7b5-kube-api-access-hhqb6\") pod \"collect-profiles-29322285-s2868\" (UID: \"72d4a425-8329-455c-8075-2c1657fec7b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.296806 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/72d4a425-8329-455c-8075-2c1657fec7b5-secret-volume\") pod \"collect-profiles-29322285-s2868\" (UID: \"72d4a425-8329-455c-8075-2c1657fec7b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.297212 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/72d4a425-8329-455c-8075-2c1657fec7b5-config-volume\") pod \"collect-profiles-29322285-s2868\" (UID: \"72d4a425-8329-455c-8075-2c1657fec7b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.398945 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/72d4a425-8329-455c-8075-2c1657fec7b5-config-volume\") pod \"collect-profiles-29322285-s2868\" (UID: \"72d4a425-8329-455c-8075-2c1657fec7b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.399239 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhqb6\" (UniqueName: \"kubernetes.io/projected/72d4a425-8329-455c-8075-2c1657fec7b5-kube-api-access-hhqb6\") pod \"collect-profiles-29322285-s2868\" (UID: \"72d4a425-8329-455c-8075-2c1657fec7b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.399331 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/72d4a425-8329-455c-8075-2c1657fec7b5-secret-volume\") pod \"collect-profiles-29322285-s2868\" (UID: \"72d4a425-8329-455c-8075-2c1657fec7b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.400349 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/72d4a425-8329-455c-8075-2c1657fec7b5-config-volume\") pod \"collect-profiles-29322285-s2868\" (UID: \"72d4a425-8329-455c-8075-2c1657fec7b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.408987 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/72d4a425-8329-455c-8075-2c1657fec7b5-secret-volume\") pod \"collect-profiles-29322285-s2868\" (UID: \"72d4a425-8329-455c-8075-2c1657fec7b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.425594 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhqb6\" (UniqueName: \"kubernetes.io/projected/72d4a425-8329-455c-8075-2c1657fec7b5-kube-api-access-hhqb6\") pod \"collect-profiles-29322285-s2868\" (UID: \"72d4a425-8329-455c-8075-2c1657fec7b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.494445 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" Oct 01 16:45:00 crc kubenswrapper[4726]: I1001 16:45:00.949397 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868"] Oct 01 16:45:01 crc kubenswrapper[4726]: I1001 16:45:01.298004 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" event={"ID":"72d4a425-8329-455c-8075-2c1657fec7b5","Type":"ContainerStarted","Data":"783c263707bd8ebe8f175bb887750707b3ac6f934c6581911b49a4b962625e23"} Oct 01 16:45:01 crc kubenswrapper[4726]: I1001 16:45:01.298385 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" event={"ID":"72d4a425-8329-455c-8075-2c1657fec7b5","Type":"ContainerStarted","Data":"5a712567234d00be2e0b4e8f97af9507e6367fd9d8c65025478ae93e51aa21f4"} Oct 01 16:45:01 crc kubenswrapper[4726]: I1001 16:45:01.320289 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" podStartSLOduration=1.320268187 podStartE2EDuration="1.320268187s" podCreationTimestamp="2025-10-01 16:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:45:01.312305156 +0000 UTC m=+2754.213857743" watchObservedRunningTime="2025-10-01 16:45:01.320268187 +0000 UTC m=+2754.221820774" Oct 01 16:45:02 crc kubenswrapper[4726]: I1001 16:45:02.310827 4726 generic.go:334] "Generic (PLEG): container finished" podID="72d4a425-8329-455c-8075-2c1657fec7b5" containerID="783c263707bd8ebe8f175bb887750707b3ac6f934c6581911b49a4b962625e23" exitCode=0 Oct 01 16:45:02 crc kubenswrapper[4726]: I1001 16:45:02.310874 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" event={"ID":"72d4a425-8329-455c-8075-2c1657fec7b5","Type":"ContainerDied","Data":"783c263707bd8ebe8f175bb887750707b3ac6f934c6581911b49a4b962625e23"} Oct 01 16:45:03 crc kubenswrapper[4726]: I1001 16:45:03.677274 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" Oct 01 16:45:03 crc kubenswrapper[4726]: I1001 16:45:03.770166 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/72d4a425-8329-455c-8075-2c1657fec7b5-config-volume\") pod \"72d4a425-8329-455c-8075-2c1657fec7b5\" (UID: \"72d4a425-8329-455c-8075-2c1657fec7b5\") " Oct 01 16:45:03 crc kubenswrapper[4726]: I1001 16:45:03.770634 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhqb6\" (UniqueName: \"kubernetes.io/projected/72d4a425-8329-455c-8075-2c1657fec7b5-kube-api-access-hhqb6\") pod \"72d4a425-8329-455c-8075-2c1657fec7b5\" (UID: \"72d4a425-8329-455c-8075-2c1657fec7b5\") " Oct 01 16:45:03 crc kubenswrapper[4726]: I1001 16:45:03.771494 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72d4a425-8329-455c-8075-2c1657fec7b5-config-volume" (OuterVolumeSpecName: "config-volume") pod "72d4a425-8329-455c-8075-2c1657fec7b5" (UID: "72d4a425-8329-455c-8075-2c1657fec7b5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:45:03 crc kubenswrapper[4726]: I1001 16:45:03.772207 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/72d4a425-8329-455c-8075-2c1657fec7b5-secret-volume\") pod \"72d4a425-8329-455c-8075-2c1657fec7b5\" (UID: \"72d4a425-8329-455c-8075-2c1657fec7b5\") " Oct 01 16:45:03 crc kubenswrapper[4726]: I1001 16:45:03.772774 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/72d4a425-8329-455c-8075-2c1657fec7b5-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:45:03 crc kubenswrapper[4726]: I1001 16:45:03.794429 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72d4a425-8329-455c-8075-2c1657fec7b5-kube-api-access-hhqb6" (OuterVolumeSpecName: "kube-api-access-hhqb6") pod "72d4a425-8329-455c-8075-2c1657fec7b5" (UID: "72d4a425-8329-455c-8075-2c1657fec7b5"). InnerVolumeSpecName "kube-api-access-hhqb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:45:03 crc kubenswrapper[4726]: I1001 16:45:03.794532 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72d4a425-8329-455c-8075-2c1657fec7b5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "72d4a425-8329-455c-8075-2c1657fec7b5" (UID: "72d4a425-8329-455c-8075-2c1657fec7b5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:45:03 crc kubenswrapper[4726]: I1001 16:45:03.874108 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/72d4a425-8329-455c-8075-2c1657fec7b5-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:45:03 crc kubenswrapper[4726]: I1001 16:45:03.874137 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhqb6\" (UniqueName: \"kubernetes.io/projected/72d4a425-8329-455c-8075-2c1657fec7b5-kube-api-access-hhqb6\") on node \"crc\" DevicePath \"\"" Oct 01 16:45:04 crc kubenswrapper[4726]: I1001 16:45:04.333240 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" event={"ID":"72d4a425-8329-455c-8075-2c1657fec7b5","Type":"ContainerDied","Data":"5a712567234d00be2e0b4e8f97af9507e6367fd9d8c65025478ae93e51aa21f4"} Oct 01 16:45:04 crc kubenswrapper[4726]: I1001 16:45:04.333276 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a712567234d00be2e0b4e8f97af9507e6367fd9d8c65025478ae93e51aa21f4" Oct 01 16:45:04 crc kubenswrapper[4726]: I1001 16:45:04.333280 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-s2868" Oct 01 16:45:04 crc kubenswrapper[4726]: I1001 16:45:04.387457 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4"] Oct 01 16:45:04 crc kubenswrapper[4726]: I1001 16:45:04.396840 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322240-5zpj4"] Oct 01 16:45:05 crc kubenswrapper[4726]: I1001 16:45:05.822359 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8cbbf14-6dac-425e-9cb6-8579f7a150ee" path="/var/lib/kubelet/pods/f8cbbf14-6dac-425e-9cb6-8579f7a150ee/volumes" Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.538305 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zs7mm"] Oct 01 16:45:21 crc kubenswrapper[4726]: E1001 16:45:21.539375 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d4a425-8329-455c-8075-2c1657fec7b5" containerName="collect-profiles" Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.539392 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d4a425-8329-455c-8075-2c1657fec7b5" containerName="collect-profiles" Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.539626 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="72d4a425-8329-455c-8075-2c1657fec7b5" containerName="collect-profiles" Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.541406 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.547247 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zs7mm"] Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.649289 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a5db0df-e838-4fbd-bf4b-c05220df7697-catalog-content\") pod \"certified-operators-zs7mm\" (UID: \"8a5db0df-e838-4fbd-bf4b-c05220df7697\") " pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.649397 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgc59\" (UniqueName: \"kubernetes.io/projected/8a5db0df-e838-4fbd-bf4b-c05220df7697-kube-api-access-hgc59\") pod \"certified-operators-zs7mm\" (UID: \"8a5db0df-e838-4fbd-bf4b-c05220df7697\") " pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.649471 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a5db0df-e838-4fbd-bf4b-c05220df7697-utilities\") pod \"certified-operators-zs7mm\" (UID: \"8a5db0df-e838-4fbd-bf4b-c05220df7697\") " pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.751538 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgc59\" (UniqueName: \"kubernetes.io/projected/8a5db0df-e838-4fbd-bf4b-c05220df7697-kube-api-access-hgc59\") pod \"certified-operators-zs7mm\" (UID: \"8a5db0df-e838-4fbd-bf4b-c05220df7697\") " pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.751587 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a5db0df-e838-4fbd-bf4b-c05220df7697-utilities\") pod \"certified-operators-zs7mm\" (UID: \"8a5db0df-e838-4fbd-bf4b-c05220df7697\") " pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.751687 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a5db0df-e838-4fbd-bf4b-c05220df7697-catalog-content\") pod \"certified-operators-zs7mm\" (UID: \"8a5db0df-e838-4fbd-bf4b-c05220df7697\") " pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.752242 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a5db0df-e838-4fbd-bf4b-c05220df7697-catalog-content\") pod \"certified-operators-zs7mm\" (UID: \"8a5db0df-e838-4fbd-bf4b-c05220df7697\") " pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.752904 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a5db0df-e838-4fbd-bf4b-c05220df7697-utilities\") pod \"certified-operators-zs7mm\" (UID: \"8a5db0df-e838-4fbd-bf4b-c05220df7697\") " pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.773907 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgc59\" (UniqueName: \"kubernetes.io/projected/8a5db0df-e838-4fbd-bf4b-c05220df7697-kube-api-access-hgc59\") pod \"certified-operators-zs7mm\" (UID: \"8a5db0df-e838-4fbd-bf4b-c05220df7697\") " pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:21 crc kubenswrapper[4726]: I1001 16:45:21.909929 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:22 crc kubenswrapper[4726]: I1001 16:45:22.442811 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zs7mm"] Oct 01 16:45:22 crc kubenswrapper[4726]: I1001 16:45:22.522946 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zs7mm" event={"ID":"8a5db0df-e838-4fbd-bf4b-c05220df7697","Type":"ContainerStarted","Data":"c540a213e32f9dee9f3471b8c5d6807283ccb715440fd6d3a72a6d4c8a4a7b61"} Oct 01 16:45:23 crc kubenswrapper[4726]: I1001 16:45:23.413594 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:45:23 crc kubenswrapper[4726]: I1001 16:45:23.413974 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:45:23 crc kubenswrapper[4726]: I1001 16:45:23.414037 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 16:45:23 crc kubenswrapper[4726]: I1001 16:45:23.415224 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"58493af1875a9a504d29f76fe2259664f298bbadd22e008df04c6a644f787e87"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:45:23 crc kubenswrapper[4726]: I1001 16:45:23.415333 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://58493af1875a9a504d29f76fe2259664f298bbadd22e008df04c6a644f787e87" gracePeriod=600 Oct 01 16:45:23 crc kubenswrapper[4726]: I1001 16:45:23.538876 4726 generic.go:334] "Generic (PLEG): container finished" podID="8a5db0df-e838-4fbd-bf4b-c05220df7697" containerID="170f1403a4e86d0ab73d56e76a512895393fce345c45694407e31c44b9d832cb" exitCode=0 Oct 01 16:45:23 crc kubenswrapper[4726]: I1001 16:45:23.538923 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zs7mm" event={"ID":"8a5db0df-e838-4fbd-bf4b-c05220df7697","Type":"ContainerDied","Data":"170f1403a4e86d0ab73d56e76a512895393fce345c45694407e31c44b9d832cb"} Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.267655 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.276318 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.282468 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.282605 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.282767 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.283035 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-t9dds" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.306014 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.404495 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.404568 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/261dc564-2126-415f-a5ce-a988af8a053c-config-data\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.404593 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.404620 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/261dc564-2126-415f-a5ce-a988af8a053c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.404645 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmlxv\" (UniqueName: \"kubernetes.io/projected/261dc564-2126-415f-a5ce-a988af8a053c-kube-api-access-tmlxv\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.404670 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/261dc564-2126-415f-a5ce-a988af8a053c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.404688 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.404714 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/261dc564-2126-415f-a5ce-a988af8a053c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.404741 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.506553 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.506796 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.506849 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/261dc564-2126-415f-a5ce-a988af8a053c-config-data\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.506905 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.506938 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/261dc564-2126-415f-a5ce-a988af8a053c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.506991 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmlxv\" (UniqueName: \"kubernetes.io/projected/261dc564-2126-415f-a5ce-a988af8a053c-kube-api-access-tmlxv\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.507065 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/261dc564-2126-415f-a5ce-a988af8a053c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.507093 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.507151 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/261dc564-2126-415f-a5ce-a988af8a053c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.507413 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.507773 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/261dc564-2126-415f-a5ce-a988af8a053c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.508650 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/261dc564-2126-415f-a5ce-a988af8a053c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.509092 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/261dc564-2126-415f-a5ce-a988af8a053c-config-data\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.509326 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/261dc564-2126-415f-a5ce-a988af8a053c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.513843 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.514172 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.514389 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.521806 4726 scope.go:117] "RemoveContainer" containerID="821ee00b78fb3620fef1e00d43d963d218a9f1d739df5183f5fd83421f603b7e" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.526642 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmlxv\" (UniqueName: \"kubernetes.io/projected/261dc564-2126-415f-a5ce-a988af8a053c-kube-api-access-tmlxv\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.548665 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " pod="openstack/tempest-tests-tempest" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.555270 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="58493af1875a9a504d29f76fe2259664f298bbadd22e008df04c6a644f787e87" exitCode=0 Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.555340 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"58493af1875a9a504d29f76fe2259664f298bbadd22e008df04c6a644f787e87"} Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.555369 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636"} Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.555385 4726 scope.go:117] "RemoveContainer" containerID="d5db0391a012e697dd720c793349dccaaea192bfcf2cd1866defdd1ff43f993c" Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.563652 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zs7mm" event={"ID":"8a5db0df-e838-4fbd-bf4b-c05220df7697","Type":"ContainerStarted","Data":"43b291c4fe0486939468a5aecf0e74ddde2c7c0ea4f224c5adfe3d0a0c2d7349"} Oct 01 16:45:24 crc kubenswrapper[4726]: I1001 16:45:24.596644 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 16:45:25 crc kubenswrapper[4726]: W1001 16:45:25.094381 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod261dc564_2126_415f_a5ce_a988af8a053c.slice/crio-265510c1f80135e34f9959c1824ad7a905af7cbcbf00cdef7d2ec67c76500592 WatchSource:0}: Error finding container 265510c1f80135e34f9959c1824ad7a905af7cbcbf00cdef7d2ec67c76500592: Status 404 returned error can't find the container with id 265510c1f80135e34f9959c1824ad7a905af7cbcbf00cdef7d2ec67c76500592 Oct 01 16:45:25 crc kubenswrapper[4726]: I1001 16:45:25.097417 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 01 16:45:25 crc kubenswrapper[4726]: I1001 16:45:25.581963 4726 generic.go:334] "Generic (PLEG): container finished" podID="8a5db0df-e838-4fbd-bf4b-c05220df7697" containerID="43b291c4fe0486939468a5aecf0e74ddde2c7c0ea4f224c5adfe3d0a0c2d7349" exitCode=0 Oct 01 16:45:25 crc kubenswrapper[4726]: I1001 16:45:25.582030 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zs7mm" event={"ID":"8a5db0df-e838-4fbd-bf4b-c05220df7697","Type":"ContainerDied","Data":"43b291c4fe0486939468a5aecf0e74ddde2c7c0ea4f224c5adfe3d0a0c2d7349"} Oct 01 16:45:25 crc kubenswrapper[4726]: I1001 16:45:25.591360 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"261dc564-2126-415f-a5ce-a988af8a053c","Type":"ContainerStarted","Data":"265510c1f80135e34f9959c1824ad7a905af7cbcbf00cdef7d2ec67c76500592"} Oct 01 16:45:26 crc kubenswrapper[4726]: I1001 16:45:26.604448 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zs7mm" event={"ID":"8a5db0df-e838-4fbd-bf4b-c05220df7697","Type":"ContainerStarted","Data":"a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4"} Oct 01 16:45:26 crc kubenswrapper[4726]: I1001 16:45:26.634503 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zs7mm" podStartSLOduration=3.136398085 podStartE2EDuration="5.634482568s" podCreationTimestamp="2025-10-01 16:45:21 +0000 UTC" firstStartedPulling="2025-10-01 16:45:23.541824781 +0000 UTC m=+2776.443377368" lastFinishedPulling="2025-10-01 16:45:26.039909264 +0000 UTC m=+2778.941461851" observedRunningTime="2025-10-01 16:45:26.623953023 +0000 UTC m=+2779.525505600" watchObservedRunningTime="2025-10-01 16:45:26.634482568 +0000 UTC m=+2779.536035145" Oct 01 16:45:31 crc kubenswrapper[4726]: I1001 16:45:31.910718 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:31 crc kubenswrapper[4726]: I1001 16:45:31.911248 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:31 crc kubenswrapper[4726]: I1001 16:45:31.965235 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:32 crc kubenswrapper[4726]: I1001 16:45:32.764785 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:45:32 crc kubenswrapper[4726]: I1001 16:45:32.823235 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zs7mm"] Oct 01 16:45:34 crc kubenswrapper[4726]: I1001 16:45:34.709584 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zs7mm" podUID="8a5db0df-e838-4fbd-bf4b-c05220df7697" containerName="registry-server" containerID="cri-o://a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4" gracePeriod=2 Oct 01 16:45:35 crc kubenswrapper[4726]: I1001 16:45:35.730977 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zs7mm" event={"ID":"8a5db0df-e838-4fbd-bf4b-c05220df7697","Type":"ContainerDied","Data":"a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4"} Oct 01 16:45:35 crc kubenswrapper[4726]: I1001 16:45:35.730887 4726 generic.go:334] "Generic (PLEG): container finished" podID="8a5db0df-e838-4fbd-bf4b-c05220df7697" containerID="a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4" exitCode=0 Oct 01 16:45:41 crc kubenswrapper[4726]: E1001 16:45:41.911798 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4 is running failed: container process not found" containerID="a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 16:45:41 crc kubenswrapper[4726]: E1001 16:45:41.913338 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4 is running failed: container process not found" containerID="a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 16:45:41 crc kubenswrapper[4726]: E1001 16:45:41.913604 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4 is running failed: container process not found" containerID="a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 16:45:41 crc kubenswrapper[4726]: E1001 16:45:41.913637 4726 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-zs7mm" podUID="8a5db0df-e838-4fbd-bf4b-c05220df7697" containerName="registry-server" Oct 01 16:45:49 crc kubenswrapper[4726]: I1001 16:45:49.474137 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v4nwq"] Oct 01 16:45:49 crc kubenswrapper[4726]: I1001 16:45:49.476462 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:45:49 crc kubenswrapper[4726]: I1001 16:45:49.490204 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4nwq"] Oct 01 16:45:49 crc kubenswrapper[4726]: I1001 16:45:49.548803 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-catalog-content\") pod \"redhat-marketplace-v4nwq\" (UID: \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\") " pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:45:49 crc kubenswrapper[4726]: I1001 16:45:49.548891 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-utilities\") pod \"redhat-marketplace-v4nwq\" (UID: \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\") " pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:45:49 crc kubenswrapper[4726]: I1001 16:45:49.548930 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc7pq\" (UniqueName: \"kubernetes.io/projected/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-kube-api-access-bc7pq\") pod \"redhat-marketplace-v4nwq\" (UID: \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\") " pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:45:49 crc kubenswrapper[4726]: I1001 16:45:49.650785 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-catalog-content\") pod \"redhat-marketplace-v4nwq\" (UID: \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\") " pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:45:49 crc kubenswrapper[4726]: I1001 16:45:49.650886 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-utilities\") pod \"redhat-marketplace-v4nwq\" (UID: \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\") " pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:45:49 crc kubenswrapper[4726]: I1001 16:45:49.650927 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc7pq\" (UniqueName: \"kubernetes.io/projected/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-kube-api-access-bc7pq\") pod \"redhat-marketplace-v4nwq\" (UID: \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\") " pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:45:49 crc kubenswrapper[4726]: I1001 16:45:49.651732 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-catalog-content\") pod \"redhat-marketplace-v4nwq\" (UID: \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\") " pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:45:49 crc kubenswrapper[4726]: I1001 16:45:49.656287 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-utilities\") pod \"redhat-marketplace-v4nwq\" (UID: \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\") " pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:45:49 crc kubenswrapper[4726]: I1001 16:45:49.695280 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc7pq\" (UniqueName: \"kubernetes.io/projected/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-kube-api-access-bc7pq\") pod \"redhat-marketplace-v4nwq\" (UID: \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\") " pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:45:49 crc kubenswrapper[4726]: I1001 16:45:49.809467 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:45:51 crc kubenswrapper[4726]: E1001 16:45:51.911774 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4 is running failed: container process not found" containerID="a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 16:45:51 crc kubenswrapper[4726]: E1001 16:45:51.913134 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4 is running failed: container process not found" containerID="a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 16:45:51 crc kubenswrapper[4726]: E1001 16:45:51.913878 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4 is running failed: container process not found" containerID="a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 16:45:51 crc kubenswrapper[4726]: E1001 16:45:51.913948 4726 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-zs7mm" podUID="8a5db0df-e838-4fbd-bf4b-c05220df7697" containerName="registry-server" Oct 01 16:45:59 crc kubenswrapper[4726]: E1001 16:45:59.935817 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 01 16:45:59 crc kubenswrapper[4726]: E1001 16:45:59.936436 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tmlxv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(261dc564-2126-415f-a5ce-a988af8a053c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:45:59 crc kubenswrapper[4726]: E1001 16:45:59.937977 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="261dc564-2126-415f-a5ce-a988af8a053c" Oct 01 16:45:59 crc kubenswrapper[4726]: I1001 16:45:59.959210 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.005731 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zs7mm" Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.005955 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zs7mm" event={"ID":"8a5db0df-e838-4fbd-bf4b-c05220df7697","Type":"ContainerDied","Data":"c540a213e32f9dee9f3471b8c5d6807283ccb715440fd6d3a72a6d4c8a4a7b61"} Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.006020 4726 scope.go:117] "RemoveContainer" containerID="a6304c490ab138acf56a8f83b37eddb5db7df0962fa652e5b685efeaa5ad48f4" Oct 01 16:46:00 crc kubenswrapper[4726]: E1001 16:46:00.007660 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="261dc564-2126-415f-a5ce-a988af8a053c" Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.047881 4726 scope.go:117] "RemoveContainer" containerID="43b291c4fe0486939468a5aecf0e74ddde2c7c0ea4f224c5adfe3d0a0c2d7349" Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.072273 4726 scope.go:117] "RemoveContainer" containerID="170f1403a4e86d0ab73d56e76a512895393fce345c45694407e31c44b9d832cb" Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.082142 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a5db0df-e838-4fbd-bf4b-c05220df7697-catalog-content\") pod \"8a5db0df-e838-4fbd-bf4b-c05220df7697\" (UID: \"8a5db0df-e838-4fbd-bf4b-c05220df7697\") " Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.082376 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgc59\" (UniqueName: \"kubernetes.io/projected/8a5db0df-e838-4fbd-bf4b-c05220df7697-kube-api-access-hgc59\") pod \"8a5db0df-e838-4fbd-bf4b-c05220df7697\" (UID: \"8a5db0df-e838-4fbd-bf4b-c05220df7697\") " Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.082419 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a5db0df-e838-4fbd-bf4b-c05220df7697-utilities\") pod \"8a5db0df-e838-4fbd-bf4b-c05220df7697\" (UID: \"8a5db0df-e838-4fbd-bf4b-c05220df7697\") " Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.084761 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a5db0df-e838-4fbd-bf4b-c05220df7697-utilities" (OuterVolumeSpecName: "utilities") pod "8a5db0df-e838-4fbd-bf4b-c05220df7697" (UID: "8a5db0df-e838-4fbd-bf4b-c05220df7697"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.090370 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a5db0df-e838-4fbd-bf4b-c05220df7697-kube-api-access-hgc59" (OuterVolumeSpecName: "kube-api-access-hgc59") pod "8a5db0df-e838-4fbd-bf4b-c05220df7697" (UID: "8a5db0df-e838-4fbd-bf4b-c05220df7697"). InnerVolumeSpecName "kube-api-access-hgc59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.120657 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a5db0df-e838-4fbd-bf4b-c05220df7697-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a5db0df-e838-4fbd-bf4b-c05220df7697" (UID: "8a5db0df-e838-4fbd-bf4b-c05220df7697"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.188369 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgc59\" (UniqueName: \"kubernetes.io/projected/8a5db0df-e838-4fbd-bf4b-c05220df7697-kube-api-access-hgc59\") on node \"crc\" DevicePath \"\"" Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.188725 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a5db0df-e838-4fbd-bf4b-c05220df7697-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.188822 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a5db0df-e838-4fbd-bf4b-c05220df7697-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.339794 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zs7mm"] Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.347207 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zs7mm"] Oct 01 16:46:00 crc kubenswrapper[4726]: I1001 16:46:00.402555 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4nwq"] Oct 01 16:46:00 crc kubenswrapper[4726]: W1001 16:46:00.409539 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cc4aad3_4cfe_4298_82f5_4dabb48b11c4.slice/crio-73454f89db6d75801f367fdc7befebf2c79b1179a45c65b935d94edce8554a72 WatchSource:0}: Error finding container 73454f89db6d75801f367fdc7befebf2c79b1179a45c65b935d94edce8554a72: Status 404 returned error can't find the container with id 73454f89db6d75801f367fdc7befebf2c79b1179a45c65b935d94edce8554a72 Oct 01 16:46:01 crc kubenswrapper[4726]: I1001 16:46:01.018971 4726 generic.go:334] "Generic (PLEG): container finished" podID="2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" containerID="f3f0ff7bac0808631441c571fc2a05a3be8bf2ab88e0c25f06eb86e72bc08c1b" exitCode=0 Oct 01 16:46:01 crc kubenswrapper[4726]: I1001 16:46:01.019098 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4nwq" event={"ID":"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4","Type":"ContainerDied","Data":"f3f0ff7bac0808631441c571fc2a05a3be8bf2ab88e0c25f06eb86e72bc08c1b"} Oct 01 16:46:01 crc kubenswrapper[4726]: I1001 16:46:01.019294 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4nwq" event={"ID":"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4","Type":"ContainerStarted","Data":"73454f89db6d75801f367fdc7befebf2c79b1179a45c65b935d94edce8554a72"} Oct 01 16:46:01 crc kubenswrapper[4726]: I1001 16:46:01.820639 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a5db0df-e838-4fbd-bf4b-c05220df7697" path="/var/lib/kubelet/pods/8a5db0df-e838-4fbd-bf4b-c05220df7697/volumes" Oct 01 16:46:03 crc kubenswrapper[4726]: I1001 16:46:03.044717 4726 generic.go:334] "Generic (PLEG): container finished" podID="2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" containerID="c35a8da89a6b7019c45df707450f03fad38a97f08c1b3d4b6e494141c574b2d0" exitCode=0 Oct 01 16:46:03 crc kubenswrapper[4726]: I1001 16:46:03.044928 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4nwq" event={"ID":"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4","Type":"ContainerDied","Data":"c35a8da89a6b7019c45df707450f03fad38a97f08c1b3d4b6e494141c574b2d0"} Oct 01 16:46:05 crc kubenswrapper[4726]: I1001 16:46:05.074111 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4nwq" event={"ID":"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4","Type":"ContainerStarted","Data":"3473e9eb6f6d2c3c61baea0e6946e828c3703dca35b0d07ae5d8bc093de91dbe"} Oct 01 16:46:05 crc kubenswrapper[4726]: I1001 16:46:05.107965 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v4nwq" podStartSLOduration=13.226389383 podStartE2EDuration="16.107936967s" podCreationTimestamp="2025-10-01 16:45:49 +0000 UTC" firstStartedPulling="2025-10-01 16:46:01.023959989 +0000 UTC m=+2813.925512566" lastFinishedPulling="2025-10-01 16:46:03.905507573 +0000 UTC m=+2816.807060150" observedRunningTime="2025-10-01 16:46:05.097899906 +0000 UTC m=+2817.999452483" watchObservedRunningTime="2025-10-01 16:46:05.107936967 +0000 UTC m=+2818.009489584" Oct 01 16:46:09 crc kubenswrapper[4726]: I1001 16:46:09.820421 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:46:09 crc kubenswrapper[4726]: I1001 16:46:09.821180 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:46:09 crc kubenswrapper[4726]: I1001 16:46:09.863199 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:46:10 crc kubenswrapper[4726]: I1001 16:46:10.195857 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:46:10 crc kubenswrapper[4726]: I1001 16:46:10.263442 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4nwq"] Oct 01 16:46:12 crc kubenswrapper[4726]: I1001 16:46:12.148289 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v4nwq" podUID="2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" containerName="registry-server" containerID="cri-o://3473e9eb6f6d2c3c61baea0e6946e828c3703dca35b0d07ae5d8bc093de91dbe" gracePeriod=2 Oct 01 16:46:12 crc kubenswrapper[4726]: I1001 16:46:12.630644 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:46:12 crc kubenswrapper[4726]: I1001 16:46:12.750572 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-catalog-content\") pod \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\" (UID: \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\") " Oct 01 16:46:12 crc kubenswrapper[4726]: I1001 16:46:12.750674 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bc7pq\" (UniqueName: \"kubernetes.io/projected/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-kube-api-access-bc7pq\") pod \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\" (UID: \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\") " Oct 01 16:46:12 crc kubenswrapper[4726]: I1001 16:46:12.750821 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-utilities\") pod \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\" (UID: \"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4\") " Oct 01 16:46:12 crc kubenswrapper[4726]: I1001 16:46:12.751935 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-utilities" (OuterVolumeSpecName: "utilities") pod "2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" (UID: "2cc4aad3-4cfe-4298-82f5-4dabb48b11c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:46:12 crc kubenswrapper[4726]: I1001 16:46:12.757002 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-kube-api-access-bc7pq" (OuterVolumeSpecName: "kube-api-access-bc7pq") pod "2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" (UID: "2cc4aad3-4cfe-4298-82f5-4dabb48b11c4"). InnerVolumeSpecName "kube-api-access-bc7pq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:46:12 crc kubenswrapper[4726]: I1001 16:46:12.766986 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" (UID: "2cc4aad3-4cfe-4298-82f5-4dabb48b11c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:46:12 crc kubenswrapper[4726]: I1001 16:46:12.852457 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:46:12 crc kubenswrapper[4726]: I1001 16:46:12.852490 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:46:12 crc kubenswrapper[4726]: I1001 16:46:12.852501 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bc7pq\" (UniqueName: \"kubernetes.io/projected/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4-kube-api-access-bc7pq\") on node \"crc\" DevicePath \"\"" Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.158024 4726 generic.go:334] "Generic (PLEG): container finished" podID="2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" containerID="3473e9eb6f6d2c3c61baea0e6946e828c3703dca35b0d07ae5d8bc093de91dbe" exitCode=0 Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.158122 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4nwq" event={"ID":"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4","Type":"ContainerDied","Data":"3473e9eb6f6d2c3c61baea0e6946e828c3703dca35b0d07ae5d8bc093de91dbe"} Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.158171 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v4nwq" Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.158207 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4nwq" event={"ID":"2cc4aad3-4cfe-4298-82f5-4dabb48b11c4","Type":"ContainerDied","Data":"73454f89db6d75801f367fdc7befebf2c79b1179a45c65b935d94edce8554a72"} Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.158246 4726 scope.go:117] "RemoveContainer" containerID="3473e9eb6f6d2c3c61baea0e6946e828c3703dca35b0d07ae5d8bc093de91dbe" Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.178039 4726 scope.go:117] "RemoveContainer" containerID="c35a8da89a6b7019c45df707450f03fad38a97f08c1b3d4b6e494141c574b2d0" Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.207288 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4nwq"] Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.221022 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4nwq"] Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.224354 4726 scope.go:117] "RemoveContainer" containerID="f3f0ff7bac0808631441c571fc2a05a3be8bf2ab88e0c25f06eb86e72bc08c1b" Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.302135 4726 scope.go:117] "RemoveContainer" containerID="3473e9eb6f6d2c3c61baea0e6946e828c3703dca35b0d07ae5d8bc093de91dbe" Oct 01 16:46:13 crc kubenswrapper[4726]: E1001 16:46:13.302814 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3473e9eb6f6d2c3c61baea0e6946e828c3703dca35b0d07ae5d8bc093de91dbe\": container with ID starting with 3473e9eb6f6d2c3c61baea0e6946e828c3703dca35b0d07ae5d8bc093de91dbe not found: ID does not exist" containerID="3473e9eb6f6d2c3c61baea0e6946e828c3703dca35b0d07ae5d8bc093de91dbe" Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.302869 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3473e9eb6f6d2c3c61baea0e6946e828c3703dca35b0d07ae5d8bc093de91dbe"} err="failed to get container status \"3473e9eb6f6d2c3c61baea0e6946e828c3703dca35b0d07ae5d8bc093de91dbe\": rpc error: code = NotFound desc = could not find container \"3473e9eb6f6d2c3c61baea0e6946e828c3703dca35b0d07ae5d8bc093de91dbe\": container with ID starting with 3473e9eb6f6d2c3c61baea0e6946e828c3703dca35b0d07ae5d8bc093de91dbe not found: ID does not exist" Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.302902 4726 scope.go:117] "RemoveContainer" containerID="c35a8da89a6b7019c45df707450f03fad38a97f08c1b3d4b6e494141c574b2d0" Oct 01 16:46:13 crc kubenswrapper[4726]: E1001 16:46:13.303305 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c35a8da89a6b7019c45df707450f03fad38a97f08c1b3d4b6e494141c574b2d0\": container with ID starting with c35a8da89a6b7019c45df707450f03fad38a97f08c1b3d4b6e494141c574b2d0 not found: ID does not exist" containerID="c35a8da89a6b7019c45df707450f03fad38a97f08c1b3d4b6e494141c574b2d0" Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.303354 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c35a8da89a6b7019c45df707450f03fad38a97f08c1b3d4b6e494141c574b2d0"} err="failed to get container status \"c35a8da89a6b7019c45df707450f03fad38a97f08c1b3d4b6e494141c574b2d0\": rpc error: code = NotFound desc = could not find container \"c35a8da89a6b7019c45df707450f03fad38a97f08c1b3d4b6e494141c574b2d0\": container with ID starting with c35a8da89a6b7019c45df707450f03fad38a97f08c1b3d4b6e494141c574b2d0 not found: ID does not exist" Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.303391 4726 scope.go:117] "RemoveContainer" containerID="f3f0ff7bac0808631441c571fc2a05a3be8bf2ab88e0c25f06eb86e72bc08c1b" Oct 01 16:46:13 crc kubenswrapper[4726]: E1001 16:46:13.303918 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3f0ff7bac0808631441c571fc2a05a3be8bf2ab88e0c25f06eb86e72bc08c1b\": container with ID starting with f3f0ff7bac0808631441c571fc2a05a3be8bf2ab88e0c25f06eb86e72bc08c1b not found: ID does not exist" containerID="f3f0ff7bac0808631441c571fc2a05a3be8bf2ab88e0c25f06eb86e72bc08c1b" Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.303950 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3f0ff7bac0808631441c571fc2a05a3be8bf2ab88e0c25f06eb86e72bc08c1b"} err="failed to get container status \"f3f0ff7bac0808631441c571fc2a05a3be8bf2ab88e0c25f06eb86e72bc08c1b\": rpc error: code = NotFound desc = could not find container \"f3f0ff7bac0808631441c571fc2a05a3be8bf2ab88e0c25f06eb86e72bc08c1b\": container with ID starting with f3f0ff7bac0808631441c571fc2a05a3be8bf2ab88e0c25f06eb86e72bc08c1b not found: ID does not exist" Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.306546 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 01 16:46:13 crc kubenswrapper[4726]: I1001 16:46:13.820657 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" path="/var/lib/kubelet/pods/2cc4aad3-4cfe-4298-82f5-4dabb48b11c4/volumes" Oct 01 16:46:15 crc kubenswrapper[4726]: I1001 16:46:15.190442 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"261dc564-2126-415f-a5ce-a988af8a053c","Type":"ContainerStarted","Data":"2b9050298664a40b33ab5e352d7bec523157f3dd66ab03da1a7a0c5e208f08ed"} Oct 01 16:46:15 crc kubenswrapper[4726]: I1001 16:46:15.225119 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.02113149 podStartE2EDuration="52.225095545s" podCreationTimestamp="2025-10-01 16:45:23 +0000 UTC" firstStartedPulling="2025-10-01 16:45:25.09990771 +0000 UTC m=+2778.001460297" lastFinishedPulling="2025-10-01 16:46:13.303871775 +0000 UTC m=+2826.205424352" observedRunningTime="2025-10-01 16:46:15.223002645 +0000 UTC m=+2828.124555282" watchObservedRunningTime="2025-10-01 16:46:15.225095545 +0000 UTC m=+2828.126648132" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.069761 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f5x77"] Oct 01 16:47:05 crc kubenswrapper[4726]: E1001 16:47:05.070788 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" containerName="registry-server" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.070804 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" containerName="registry-server" Oct 01 16:47:05 crc kubenswrapper[4726]: E1001 16:47:05.070828 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5db0df-e838-4fbd-bf4b-c05220df7697" containerName="registry-server" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.070836 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5db0df-e838-4fbd-bf4b-c05220df7697" containerName="registry-server" Oct 01 16:47:05 crc kubenswrapper[4726]: E1001 16:47:05.070855 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" containerName="extract-utilities" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.070866 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" containerName="extract-utilities" Oct 01 16:47:05 crc kubenswrapper[4726]: E1001 16:47:05.070890 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5db0df-e838-4fbd-bf4b-c05220df7697" containerName="extract-content" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.070898 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5db0df-e838-4fbd-bf4b-c05220df7697" containerName="extract-content" Oct 01 16:47:05 crc kubenswrapper[4726]: E1001 16:47:05.070928 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" containerName="extract-content" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.070935 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" containerName="extract-content" Oct 01 16:47:05 crc kubenswrapper[4726]: E1001 16:47:05.070944 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5db0df-e838-4fbd-bf4b-c05220df7697" containerName="extract-utilities" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.070951 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5db0df-e838-4fbd-bf4b-c05220df7697" containerName="extract-utilities" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.071221 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cc4aad3-4cfe-4298-82f5-4dabb48b11c4" containerName="registry-server" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.071252 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a5db0df-e838-4fbd-bf4b-c05220df7697" containerName="registry-server" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.072999 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.094352 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f5x77"] Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.149322 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxw7c\" (UniqueName: \"kubernetes.io/projected/8d1dd34c-02b0-44f5-954e-63576898f64f-kube-api-access-nxw7c\") pod \"community-operators-f5x77\" (UID: \"8d1dd34c-02b0-44f5-954e-63576898f64f\") " pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.149571 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1dd34c-02b0-44f5-954e-63576898f64f-utilities\") pod \"community-operators-f5x77\" (UID: \"8d1dd34c-02b0-44f5-954e-63576898f64f\") " pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.149738 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1dd34c-02b0-44f5-954e-63576898f64f-catalog-content\") pod \"community-operators-f5x77\" (UID: \"8d1dd34c-02b0-44f5-954e-63576898f64f\") " pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.252075 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxw7c\" (UniqueName: \"kubernetes.io/projected/8d1dd34c-02b0-44f5-954e-63576898f64f-kube-api-access-nxw7c\") pod \"community-operators-f5x77\" (UID: \"8d1dd34c-02b0-44f5-954e-63576898f64f\") " pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.252245 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1dd34c-02b0-44f5-954e-63576898f64f-utilities\") pod \"community-operators-f5x77\" (UID: \"8d1dd34c-02b0-44f5-954e-63576898f64f\") " pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.252291 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1dd34c-02b0-44f5-954e-63576898f64f-catalog-content\") pod \"community-operators-f5x77\" (UID: \"8d1dd34c-02b0-44f5-954e-63576898f64f\") " pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.252813 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1dd34c-02b0-44f5-954e-63576898f64f-utilities\") pod \"community-operators-f5x77\" (UID: \"8d1dd34c-02b0-44f5-954e-63576898f64f\") " pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.252860 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1dd34c-02b0-44f5-954e-63576898f64f-catalog-content\") pod \"community-operators-f5x77\" (UID: \"8d1dd34c-02b0-44f5-954e-63576898f64f\") " pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.279634 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxw7c\" (UniqueName: \"kubernetes.io/projected/8d1dd34c-02b0-44f5-954e-63576898f64f-kube-api-access-nxw7c\") pod \"community-operators-f5x77\" (UID: \"8d1dd34c-02b0-44f5-954e-63576898f64f\") " pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.395598 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:05 crc kubenswrapper[4726]: I1001 16:47:05.960152 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f5x77"] Oct 01 16:47:06 crc kubenswrapper[4726]: I1001 16:47:06.768726 4726 generic.go:334] "Generic (PLEG): container finished" podID="8d1dd34c-02b0-44f5-954e-63576898f64f" containerID="8756729b2a31a1a7aa03c9f432a6179ed0813e62854fffe2753512b229583f6a" exitCode=0 Oct 01 16:47:06 crc kubenswrapper[4726]: I1001 16:47:06.768827 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5x77" event={"ID":"8d1dd34c-02b0-44f5-954e-63576898f64f","Type":"ContainerDied","Data":"8756729b2a31a1a7aa03c9f432a6179ed0813e62854fffe2753512b229583f6a"} Oct 01 16:47:06 crc kubenswrapper[4726]: I1001 16:47:06.769271 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5x77" event={"ID":"8d1dd34c-02b0-44f5-954e-63576898f64f","Type":"ContainerStarted","Data":"6e449c4a95edb5a1d398144c9539cbbf2df43338f1bb9483483094702a3b6c0e"} Oct 01 16:47:06 crc kubenswrapper[4726]: I1001 16:47:06.771930 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:47:08 crc kubenswrapper[4726]: I1001 16:47:08.791393 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5x77" event={"ID":"8d1dd34c-02b0-44f5-954e-63576898f64f","Type":"ContainerStarted","Data":"7ba41ae297dc7f8425e62f2ade6ffa85b0458217e2cc18a52ddaceca88918ff9"} Oct 01 16:47:09 crc kubenswrapper[4726]: I1001 16:47:09.803794 4726 generic.go:334] "Generic (PLEG): container finished" podID="8d1dd34c-02b0-44f5-954e-63576898f64f" containerID="7ba41ae297dc7f8425e62f2ade6ffa85b0458217e2cc18a52ddaceca88918ff9" exitCode=0 Oct 01 16:47:09 crc kubenswrapper[4726]: I1001 16:47:09.803905 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5x77" event={"ID":"8d1dd34c-02b0-44f5-954e-63576898f64f","Type":"ContainerDied","Data":"7ba41ae297dc7f8425e62f2ade6ffa85b0458217e2cc18a52ddaceca88918ff9"} Oct 01 16:47:10 crc kubenswrapper[4726]: I1001 16:47:10.818890 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5x77" event={"ID":"8d1dd34c-02b0-44f5-954e-63576898f64f","Type":"ContainerStarted","Data":"5d5fcc456f676eb5f496eea4519692ec8e7a7f3c492dd8f732e5884cbdb55866"} Oct 01 16:47:10 crc kubenswrapper[4726]: I1001 16:47:10.856408 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f5x77" podStartSLOduration=2.093123057 podStartE2EDuration="5.856385073s" podCreationTimestamp="2025-10-01 16:47:05 +0000 UTC" firstStartedPulling="2025-10-01 16:47:06.771642094 +0000 UTC m=+2879.673194681" lastFinishedPulling="2025-10-01 16:47:10.53490412 +0000 UTC m=+2883.436456697" observedRunningTime="2025-10-01 16:47:10.846372883 +0000 UTC m=+2883.747925470" watchObservedRunningTime="2025-10-01 16:47:10.856385073 +0000 UTC m=+2883.757937660" Oct 01 16:47:15 crc kubenswrapper[4726]: I1001 16:47:15.396142 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:15 crc kubenswrapper[4726]: I1001 16:47:15.396703 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:15 crc kubenswrapper[4726]: I1001 16:47:15.461643 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:15 crc kubenswrapper[4726]: I1001 16:47:15.929829 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:15 crc kubenswrapper[4726]: I1001 16:47:15.989697 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f5x77"] Oct 01 16:47:17 crc kubenswrapper[4726]: I1001 16:47:17.887326 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f5x77" podUID="8d1dd34c-02b0-44f5-954e-63576898f64f" containerName="registry-server" containerID="cri-o://5d5fcc456f676eb5f496eea4519692ec8e7a7f3c492dd8f732e5884cbdb55866" gracePeriod=2 Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.423191 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.494477 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxw7c\" (UniqueName: \"kubernetes.io/projected/8d1dd34c-02b0-44f5-954e-63576898f64f-kube-api-access-nxw7c\") pod \"8d1dd34c-02b0-44f5-954e-63576898f64f\" (UID: \"8d1dd34c-02b0-44f5-954e-63576898f64f\") " Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.494670 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1dd34c-02b0-44f5-954e-63576898f64f-catalog-content\") pod \"8d1dd34c-02b0-44f5-954e-63576898f64f\" (UID: \"8d1dd34c-02b0-44f5-954e-63576898f64f\") " Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.494829 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1dd34c-02b0-44f5-954e-63576898f64f-utilities\") pod \"8d1dd34c-02b0-44f5-954e-63576898f64f\" (UID: \"8d1dd34c-02b0-44f5-954e-63576898f64f\") " Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.496224 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d1dd34c-02b0-44f5-954e-63576898f64f-utilities" (OuterVolumeSpecName: "utilities") pod "8d1dd34c-02b0-44f5-954e-63576898f64f" (UID: "8d1dd34c-02b0-44f5-954e-63576898f64f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.499711 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d1dd34c-02b0-44f5-954e-63576898f64f-kube-api-access-nxw7c" (OuterVolumeSpecName: "kube-api-access-nxw7c") pod "8d1dd34c-02b0-44f5-954e-63576898f64f" (UID: "8d1dd34c-02b0-44f5-954e-63576898f64f"). InnerVolumeSpecName "kube-api-access-nxw7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.549335 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d1dd34c-02b0-44f5-954e-63576898f64f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d1dd34c-02b0-44f5-954e-63576898f64f" (UID: "8d1dd34c-02b0-44f5-954e-63576898f64f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.597564 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1dd34c-02b0-44f5-954e-63576898f64f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.597596 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1dd34c-02b0-44f5-954e-63576898f64f-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.597608 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxw7c\" (UniqueName: \"kubernetes.io/projected/8d1dd34c-02b0-44f5-954e-63576898f64f-kube-api-access-nxw7c\") on node \"crc\" DevicePath \"\"" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.902189 4726 generic.go:334] "Generic (PLEG): container finished" podID="8d1dd34c-02b0-44f5-954e-63576898f64f" containerID="5d5fcc456f676eb5f496eea4519692ec8e7a7f3c492dd8f732e5884cbdb55866" exitCode=0 Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.902250 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5x77" event={"ID":"8d1dd34c-02b0-44f5-954e-63576898f64f","Type":"ContainerDied","Data":"5d5fcc456f676eb5f496eea4519692ec8e7a7f3c492dd8f732e5884cbdb55866"} Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.902299 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5x77" event={"ID":"8d1dd34c-02b0-44f5-954e-63576898f64f","Type":"ContainerDied","Data":"6e449c4a95edb5a1d398144c9539cbbf2df43338f1bb9483483094702a3b6c0e"} Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.902341 4726 scope.go:117] "RemoveContainer" containerID="5d5fcc456f676eb5f496eea4519692ec8e7a7f3c492dd8f732e5884cbdb55866" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.903244 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5x77" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.925486 4726 scope.go:117] "RemoveContainer" containerID="7ba41ae297dc7f8425e62f2ade6ffa85b0458217e2cc18a52ddaceca88918ff9" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.964564 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f5x77"] Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.977174 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f5x77"] Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.978091 4726 scope.go:117] "RemoveContainer" containerID="8756729b2a31a1a7aa03c9f432a6179ed0813e62854fffe2753512b229583f6a" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.995469 4726 scope.go:117] "RemoveContainer" containerID="5d5fcc456f676eb5f496eea4519692ec8e7a7f3c492dd8f732e5884cbdb55866" Oct 01 16:47:18 crc kubenswrapper[4726]: E1001 16:47:18.995886 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d5fcc456f676eb5f496eea4519692ec8e7a7f3c492dd8f732e5884cbdb55866\": container with ID starting with 5d5fcc456f676eb5f496eea4519692ec8e7a7f3c492dd8f732e5884cbdb55866 not found: ID does not exist" containerID="5d5fcc456f676eb5f496eea4519692ec8e7a7f3c492dd8f732e5884cbdb55866" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.995915 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d5fcc456f676eb5f496eea4519692ec8e7a7f3c492dd8f732e5884cbdb55866"} err="failed to get container status \"5d5fcc456f676eb5f496eea4519692ec8e7a7f3c492dd8f732e5884cbdb55866\": rpc error: code = NotFound desc = could not find container \"5d5fcc456f676eb5f496eea4519692ec8e7a7f3c492dd8f732e5884cbdb55866\": container with ID starting with 5d5fcc456f676eb5f496eea4519692ec8e7a7f3c492dd8f732e5884cbdb55866 not found: ID does not exist" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.995936 4726 scope.go:117] "RemoveContainer" containerID="7ba41ae297dc7f8425e62f2ade6ffa85b0458217e2cc18a52ddaceca88918ff9" Oct 01 16:47:18 crc kubenswrapper[4726]: E1001 16:47:18.996313 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ba41ae297dc7f8425e62f2ade6ffa85b0458217e2cc18a52ddaceca88918ff9\": container with ID starting with 7ba41ae297dc7f8425e62f2ade6ffa85b0458217e2cc18a52ddaceca88918ff9 not found: ID does not exist" containerID="7ba41ae297dc7f8425e62f2ade6ffa85b0458217e2cc18a52ddaceca88918ff9" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.996330 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ba41ae297dc7f8425e62f2ade6ffa85b0458217e2cc18a52ddaceca88918ff9"} err="failed to get container status \"7ba41ae297dc7f8425e62f2ade6ffa85b0458217e2cc18a52ddaceca88918ff9\": rpc error: code = NotFound desc = could not find container \"7ba41ae297dc7f8425e62f2ade6ffa85b0458217e2cc18a52ddaceca88918ff9\": container with ID starting with 7ba41ae297dc7f8425e62f2ade6ffa85b0458217e2cc18a52ddaceca88918ff9 not found: ID does not exist" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.996342 4726 scope.go:117] "RemoveContainer" containerID="8756729b2a31a1a7aa03c9f432a6179ed0813e62854fffe2753512b229583f6a" Oct 01 16:47:18 crc kubenswrapper[4726]: E1001 16:47:18.996572 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8756729b2a31a1a7aa03c9f432a6179ed0813e62854fffe2753512b229583f6a\": container with ID starting with 8756729b2a31a1a7aa03c9f432a6179ed0813e62854fffe2753512b229583f6a not found: ID does not exist" containerID="8756729b2a31a1a7aa03c9f432a6179ed0813e62854fffe2753512b229583f6a" Oct 01 16:47:18 crc kubenswrapper[4726]: I1001 16:47:18.996588 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8756729b2a31a1a7aa03c9f432a6179ed0813e62854fffe2753512b229583f6a"} err="failed to get container status \"8756729b2a31a1a7aa03c9f432a6179ed0813e62854fffe2753512b229583f6a\": rpc error: code = NotFound desc = could not find container \"8756729b2a31a1a7aa03c9f432a6179ed0813e62854fffe2753512b229583f6a\": container with ID starting with 8756729b2a31a1a7aa03c9f432a6179ed0813e62854fffe2753512b229583f6a not found: ID does not exist" Oct 01 16:47:19 crc kubenswrapper[4726]: I1001 16:47:19.825895 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d1dd34c-02b0-44f5-954e-63576898f64f" path="/var/lib/kubelet/pods/8d1dd34c-02b0-44f5-954e-63576898f64f/volumes" Oct 01 16:47:23 crc kubenswrapper[4726]: I1001 16:47:23.414374 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:47:23 crc kubenswrapper[4726]: I1001 16:47:23.415252 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:47:53 crc kubenswrapper[4726]: I1001 16:47:53.413407 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:47:53 crc kubenswrapper[4726]: I1001 16:47:53.414095 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:48:23 crc kubenswrapper[4726]: I1001 16:48:23.414262 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:48:23 crc kubenswrapper[4726]: I1001 16:48:23.415099 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:48:23 crc kubenswrapper[4726]: I1001 16:48:23.415193 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 16:48:23 crc kubenswrapper[4726]: I1001 16:48:23.416531 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:48:23 crc kubenswrapper[4726]: I1001 16:48:23.416677 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" gracePeriod=600 Oct 01 16:48:23 crc kubenswrapper[4726]: E1001 16:48:23.669716 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:48:24 crc kubenswrapper[4726]: I1001 16:48:24.534353 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" exitCode=0 Oct 01 16:48:24 crc kubenswrapper[4726]: I1001 16:48:24.534443 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636"} Oct 01 16:48:24 crc kubenswrapper[4726]: I1001 16:48:24.534708 4726 scope.go:117] "RemoveContainer" containerID="58493af1875a9a504d29f76fe2259664f298bbadd22e008df04c6a644f787e87" Oct 01 16:48:24 crc kubenswrapper[4726]: I1001 16:48:24.535730 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:48:24 crc kubenswrapper[4726]: E1001 16:48:24.536312 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:48:39 crc kubenswrapper[4726]: I1001 16:48:39.807996 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:48:39 crc kubenswrapper[4726]: E1001 16:48:39.809537 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:48:52 crc kubenswrapper[4726]: I1001 16:48:52.808855 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:48:52 crc kubenswrapper[4726]: E1001 16:48:52.810466 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:49:04 crc kubenswrapper[4726]: I1001 16:49:04.809132 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:49:04 crc kubenswrapper[4726]: E1001 16:49:04.810005 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:49:16 crc kubenswrapper[4726]: I1001 16:49:16.808852 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:49:16 crc kubenswrapper[4726]: E1001 16:49:16.810252 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:49:31 crc kubenswrapper[4726]: I1001 16:49:31.809167 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:49:31 crc kubenswrapper[4726]: E1001 16:49:31.811572 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:49:42 crc kubenswrapper[4726]: I1001 16:49:42.810457 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:49:42 crc kubenswrapper[4726]: E1001 16:49:42.811273 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:49:54 crc kubenswrapper[4726]: I1001 16:49:54.808661 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:49:54 crc kubenswrapper[4726]: E1001 16:49:54.809382 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:50:07 crc kubenswrapper[4726]: I1001 16:50:07.822828 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:50:07 crc kubenswrapper[4726]: E1001 16:50:07.823889 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:50:22 crc kubenswrapper[4726]: I1001 16:50:22.808516 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:50:22 crc kubenswrapper[4726]: E1001 16:50:22.809790 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:50:37 crc kubenswrapper[4726]: I1001 16:50:37.819114 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:50:37 crc kubenswrapper[4726]: E1001 16:50:37.820374 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:50:48 crc kubenswrapper[4726]: I1001 16:50:48.807954 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:50:48 crc kubenswrapper[4726]: E1001 16:50:48.808723 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:51:00 crc kubenswrapper[4726]: I1001 16:51:00.809008 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:51:00 crc kubenswrapper[4726]: E1001 16:51:00.810373 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:51:14 crc kubenswrapper[4726]: I1001 16:51:14.809404 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:51:14 crc kubenswrapper[4726]: E1001 16:51:14.810535 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:51:26 crc kubenswrapper[4726]: I1001 16:51:26.808885 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:51:26 crc kubenswrapper[4726]: E1001 16:51:26.809991 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.396284 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pwjfw"] Oct 01 16:51:30 crc kubenswrapper[4726]: E1001 16:51:30.397812 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1dd34c-02b0-44f5-954e-63576898f64f" containerName="extract-content" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.397848 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1dd34c-02b0-44f5-954e-63576898f64f" containerName="extract-content" Oct 01 16:51:30 crc kubenswrapper[4726]: E1001 16:51:30.397913 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1dd34c-02b0-44f5-954e-63576898f64f" containerName="extract-utilities" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.397931 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1dd34c-02b0-44f5-954e-63576898f64f" containerName="extract-utilities" Oct 01 16:51:30 crc kubenswrapper[4726]: E1001 16:51:30.397976 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1dd34c-02b0-44f5-954e-63576898f64f" containerName="registry-server" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.397998 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1dd34c-02b0-44f5-954e-63576898f64f" containerName="registry-server" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.398508 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d1dd34c-02b0-44f5-954e-63576898f64f" containerName="registry-server" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.414256 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.432257 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pwjfw"] Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.465875 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1aead31-ac35-498f-b23f-cdfc25d72c88-catalog-content\") pod \"redhat-operators-pwjfw\" (UID: \"d1aead31-ac35-498f-b23f-cdfc25d72c88\") " pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.466102 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qslj2\" (UniqueName: \"kubernetes.io/projected/d1aead31-ac35-498f-b23f-cdfc25d72c88-kube-api-access-qslj2\") pod \"redhat-operators-pwjfw\" (UID: \"d1aead31-ac35-498f-b23f-cdfc25d72c88\") " pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.466820 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1aead31-ac35-498f-b23f-cdfc25d72c88-utilities\") pod \"redhat-operators-pwjfw\" (UID: \"d1aead31-ac35-498f-b23f-cdfc25d72c88\") " pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.568747 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1aead31-ac35-498f-b23f-cdfc25d72c88-catalog-content\") pod \"redhat-operators-pwjfw\" (UID: \"d1aead31-ac35-498f-b23f-cdfc25d72c88\") " pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.569036 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qslj2\" (UniqueName: \"kubernetes.io/projected/d1aead31-ac35-498f-b23f-cdfc25d72c88-kube-api-access-qslj2\") pod \"redhat-operators-pwjfw\" (UID: \"d1aead31-ac35-498f-b23f-cdfc25d72c88\") " pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.569259 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1aead31-ac35-498f-b23f-cdfc25d72c88-utilities\") pod \"redhat-operators-pwjfw\" (UID: \"d1aead31-ac35-498f-b23f-cdfc25d72c88\") " pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.569858 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1aead31-ac35-498f-b23f-cdfc25d72c88-utilities\") pod \"redhat-operators-pwjfw\" (UID: \"d1aead31-ac35-498f-b23f-cdfc25d72c88\") " pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.570230 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1aead31-ac35-498f-b23f-cdfc25d72c88-catalog-content\") pod \"redhat-operators-pwjfw\" (UID: \"d1aead31-ac35-498f-b23f-cdfc25d72c88\") " pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.595087 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qslj2\" (UniqueName: \"kubernetes.io/projected/d1aead31-ac35-498f-b23f-cdfc25d72c88-kube-api-access-qslj2\") pod \"redhat-operators-pwjfw\" (UID: \"d1aead31-ac35-498f-b23f-cdfc25d72c88\") " pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:51:30 crc kubenswrapper[4726]: I1001 16:51:30.763807 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:51:31 crc kubenswrapper[4726]: I1001 16:51:31.268326 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pwjfw"] Oct 01 16:51:31 crc kubenswrapper[4726]: I1001 16:51:31.552277 4726 generic.go:334] "Generic (PLEG): container finished" podID="d1aead31-ac35-498f-b23f-cdfc25d72c88" containerID="190b6fef6a0501c29d59acffb597e0d5cd54b2310aa997d6731fcc9f4534e864" exitCode=0 Oct 01 16:51:31 crc kubenswrapper[4726]: I1001 16:51:31.552337 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwjfw" event={"ID":"d1aead31-ac35-498f-b23f-cdfc25d72c88","Type":"ContainerDied","Data":"190b6fef6a0501c29d59acffb597e0d5cd54b2310aa997d6731fcc9f4534e864"} Oct 01 16:51:31 crc kubenswrapper[4726]: I1001 16:51:31.552404 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwjfw" event={"ID":"d1aead31-ac35-498f-b23f-cdfc25d72c88","Type":"ContainerStarted","Data":"cca681ac984e71ef24e085b59e7fdf569e92d2ebaf9c9c698a34c284cb6c4b84"} Oct 01 16:51:32 crc kubenswrapper[4726]: I1001 16:51:32.565006 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwjfw" event={"ID":"d1aead31-ac35-498f-b23f-cdfc25d72c88","Type":"ContainerStarted","Data":"394e5a68211bcf2a3333c126df01089cbe6e2c105fea1b35e7bf6c51afdc66b9"} Oct 01 16:51:36 crc kubenswrapper[4726]: I1001 16:51:36.640079 4726 generic.go:334] "Generic (PLEG): container finished" podID="d1aead31-ac35-498f-b23f-cdfc25d72c88" containerID="394e5a68211bcf2a3333c126df01089cbe6e2c105fea1b35e7bf6c51afdc66b9" exitCode=0 Oct 01 16:51:36 crc kubenswrapper[4726]: I1001 16:51:36.640150 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwjfw" event={"ID":"d1aead31-ac35-498f-b23f-cdfc25d72c88","Type":"ContainerDied","Data":"394e5a68211bcf2a3333c126df01089cbe6e2c105fea1b35e7bf6c51afdc66b9"} Oct 01 16:51:37 crc kubenswrapper[4726]: I1001 16:51:37.652504 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwjfw" event={"ID":"d1aead31-ac35-498f-b23f-cdfc25d72c88","Type":"ContainerStarted","Data":"1b53d04820b2627adef9e43c0049e05ea088a4561469db4ec64f4a63373a6005"} Oct 01 16:51:37 crc kubenswrapper[4726]: I1001 16:51:37.684337 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pwjfw" podStartSLOduration=2.002555495 podStartE2EDuration="7.684320333s" podCreationTimestamp="2025-10-01 16:51:30 +0000 UTC" firstStartedPulling="2025-10-01 16:51:31.554691505 +0000 UTC m=+3144.456244082" lastFinishedPulling="2025-10-01 16:51:37.236456343 +0000 UTC m=+3150.138008920" observedRunningTime="2025-10-01 16:51:37.678718751 +0000 UTC m=+3150.580271358" watchObservedRunningTime="2025-10-01 16:51:37.684320333 +0000 UTC m=+3150.585872910" Oct 01 16:51:40 crc kubenswrapper[4726]: I1001 16:51:40.764620 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:51:40 crc kubenswrapper[4726]: I1001 16:51:40.765172 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:51:41 crc kubenswrapper[4726]: I1001 16:51:41.809094 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:51:41 crc kubenswrapper[4726]: E1001 16:51:41.809817 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:51:41 crc kubenswrapper[4726]: I1001 16:51:41.827086 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pwjfw" podUID="d1aead31-ac35-498f-b23f-cdfc25d72c88" containerName="registry-server" probeResult="failure" output=< Oct 01 16:51:41 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Oct 01 16:51:41 crc kubenswrapper[4726]: > Oct 01 16:51:51 crc kubenswrapper[4726]: I1001 16:51:51.828537 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pwjfw" podUID="d1aead31-ac35-498f-b23f-cdfc25d72c88" containerName="registry-server" probeResult="failure" output=< Oct 01 16:51:51 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Oct 01 16:51:51 crc kubenswrapper[4726]: > Oct 01 16:51:55 crc kubenswrapper[4726]: I1001 16:51:55.808936 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:51:55 crc kubenswrapper[4726]: E1001 16:51:55.809569 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:52:00 crc kubenswrapper[4726]: I1001 16:52:00.826171 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:52:00 crc kubenswrapper[4726]: I1001 16:52:00.915465 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:52:01 crc kubenswrapper[4726]: I1001 16:52:01.582321 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pwjfw"] Oct 01 16:52:01 crc kubenswrapper[4726]: I1001 16:52:01.943127 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pwjfw" podUID="d1aead31-ac35-498f-b23f-cdfc25d72c88" containerName="registry-server" containerID="cri-o://1b53d04820b2627adef9e43c0049e05ea088a4561469db4ec64f4a63373a6005" gracePeriod=2 Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.586851 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.732925 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1aead31-ac35-498f-b23f-cdfc25d72c88-catalog-content\") pod \"d1aead31-ac35-498f-b23f-cdfc25d72c88\" (UID: \"d1aead31-ac35-498f-b23f-cdfc25d72c88\") " Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.733038 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qslj2\" (UniqueName: \"kubernetes.io/projected/d1aead31-ac35-498f-b23f-cdfc25d72c88-kube-api-access-qslj2\") pod \"d1aead31-ac35-498f-b23f-cdfc25d72c88\" (UID: \"d1aead31-ac35-498f-b23f-cdfc25d72c88\") " Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.733127 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1aead31-ac35-498f-b23f-cdfc25d72c88-utilities\") pod \"d1aead31-ac35-498f-b23f-cdfc25d72c88\" (UID: \"d1aead31-ac35-498f-b23f-cdfc25d72c88\") " Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.735640 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1aead31-ac35-498f-b23f-cdfc25d72c88-utilities" (OuterVolumeSpecName: "utilities") pod "d1aead31-ac35-498f-b23f-cdfc25d72c88" (UID: "d1aead31-ac35-498f-b23f-cdfc25d72c88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.743317 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1aead31-ac35-498f-b23f-cdfc25d72c88-kube-api-access-qslj2" (OuterVolumeSpecName: "kube-api-access-qslj2") pod "d1aead31-ac35-498f-b23f-cdfc25d72c88" (UID: "d1aead31-ac35-498f-b23f-cdfc25d72c88"). InnerVolumeSpecName "kube-api-access-qslj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.835313 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1aead31-ac35-498f-b23f-cdfc25d72c88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1aead31-ac35-498f-b23f-cdfc25d72c88" (UID: "d1aead31-ac35-498f-b23f-cdfc25d72c88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.835687 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1aead31-ac35-498f-b23f-cdfc25d72c88-catalog-content\") pod \"d1aead31-ac35-498f-b23f-cdfc25d72c88\" (UID: \"d1aead31-ac35-498f-b23f-cdfc25d72c88\") " Oct 01 16:52:02 crc kubenswrapper[4726]: W1001 16:52:02.835788 4726 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/d1aead31-ac35-498f-b23f-cdfc25d72c88/volumes/kubernetes.io~empty-dir/catalog-content Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.835802 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1aead31-ac35-498f-b23f-cdfc25d72c88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1aead31-ac35-498f-b23f-cdfc25d72c88" (UID: "d1aead31-ac35-498f-b23f-cdfc25d72c88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.836438 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1aead31-ac35-498f-b23f-cdfc25d72c88-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.836457 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1aead31-ac35-498f-b23f-cdfc25d72c88-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.836469 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qslj2\" (UniqueName: \"kubernetes.io/projected/d1aead31-ac35-498f-b23f-cdfc25d72c88-kube-api-access-qslj2\") on node \"crc\" DevicePath \"\"" Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.954757 4726 generic.go:334] "Generic (PLEG): container finished" podID="d1aead31-ac35-498f-b23f-cdfc25d72c88" containerID="1b53d04820b2627adef9e43c0049e05ea088a4561469db4ec64f4a63373a6005" exitCode=0 Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.954804 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwjfw" event={"ID":"d1aead31-ac35-498f-b23f-cdfc25d72c88","Type":"ContainerDied","Data":"1b53d04820b2627adef9e43c0049e05ea088a4561469db4ec64f4a63373a6005"} Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.954835 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwjfw" event={"ID":"d1aead31-ac35-498f-b23f-cdfc25d72c88","Type":"ContainerDied","Data":"cca681ac984e71ef24e085b59e7fdf569e92d2ebaf9c9c698a34c284cb6c4b84"} Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.954857 4726 scope.go:117] "RemoveContainer" containerID="1b53d04820b2627adef9e43c0049e05ea088a4561469db4ec64f4a63373a6005" Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.955020 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwjfw" Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.979944 4726 scope.go:117] "RemoveContainer" containerID="394e5a68211bcf2a3333c126df01089cbe6e2c105fea1b35e7bf6c51afdc66b9" Oct 01 16:52:02 crc kubenswrapper[4726]: I1001 16:52:02.998289 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pwjfw"] Oct 01 16:52:03 crc kubenswrapper[4726]: I1001 16:52:03.011187 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pwjfw"] Oct 01 16:52:03 crc kubenswrapper[4726]: I1001 16:52:03.018647 4726 scope.go:117] "RemoveContainer" containerID="190b6fef6a0501c29d59acffb597e0d5cd54b2310aa997d6731fcc9f4534e864" Oct 01 16:52:03 crc kubenswrapper[4726]: I1001 16:52:03.080230 4726 scope.go:117] "RemoveContainer" containerID="1b53d04820b2627adef9e43c0049e05ea088a4561469db4ec64f4a63373a6005" Oct 01 16:52:03 crc kubenswrapper[4726]: E1001 16:52:03.080636 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b53d04820b2627adef9e43c0049e05ea088a4561469db4ec64f4a63373a6005\": container with ID starting with 1b53d04820b2627adef9e43c0049e05ea088a4561469db4ec64f4a63373a6005 not found: ID does not exist" containerID="1b53d04820b2627adef9e43c0049e05ea088a4561469db4ec64f4a63373a6005" Oct 01 16:52:03 crc kubenswrapper[4726]: I1001 16:52:03.080690 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b53d04820b2627adef9e43c0049e05ea088a4561469db4ec64f4a63373a6005"} err="failed to get container status \"1b53d04820b2627adef9e43c0049e05ea088a4561469db4ec64f4a63373a6005\": rpc error: code = NotFound desc = could not find container \"1b53d04820b2627adef9e43c0049e05ea088a4561469db4ec64f4a63373a6005\": container with ID starting with 1b53d04820b2627adef9e43c0049e05ea088a4561469db4ec64f4a63373a6005 not found: ID does not exist" Oct 01 16:52:03 crc kubenswrapper[4726]: I1001 16:52:03.080724 4726 scope.go:117] "RemoveContainer" containerID="394e5a68211bcf2a3333c126df01089cbe6e2c105fea1b35e7bf6c51afdc66b9" Oct 01 16:52:03 crc kubenswrapper[4726]: E1001 16:52:03.081204 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"394e5a68211bcf2a3333c126df01089cbe6e2c105fea1b35e7bf6c51afdc66b9\": container with ID starting with 394e5a68211bcf2a3333c126df01089cbe6e2c105fea1b35e7bf6c51afdc66b9 not found: ID does not exist" containerID="394e5a68211bcf2a3333c126df01089cbe6e2c105fea1b35e7bf6c51afdc66b9" Oct 01 16:52:03 crc kubenswrapper[4726]: I1001 16:52:03.081236 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"394e5a68211bcf2a3333c126df01089cbe6e2c105fea1b35e7bf6c51afdc66b9"} err="failed to get container status \"394e5a68211bcf2a3333c126df01089cbe6e2c105fea1b35e7bf6c51afdc66b9\": rpc error: code = NotFound desc = could not find container \"394e5a68211bcf2a3333c126df01089cbe6e2c105fea1b35e7bf6c51afdc66b9\": container with ID starting with 394e5a68211bcf2a3333c126df01089cbe6e2c105fea1b35e7bf6c51afdc66b9 not found: ID does not exist" Oct 01 16:52:03 crc kubenswrapper[4726]: I1001 16:52:03.081261 4726 scope.go:117] "RemoveContainer" containerID="190b6fef6a0501c29d59acffb597e0d5cd54b2310aa997d6731fcc9f4534e864" Oct 01 16:52:03 crc kubenswrapper[4726]: E1001 16:52:03.081537 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"190b6fef6a0501c29d59acffb597e0d5cd54b2310aa997d6731fcc9f4534e864\": container with ID starting with 190b6fef6a0501c29d59acffb597e0d5cd54b2310aa997d6731fcc9f4534e864 not found: ID does not exist" containerID="190b6fef6a0501c29d59acffb597e0d5cd54b2310aa997d6731fcc9f4534e864" Oct 01 16:52:03 crc kubenswrapper[4726]: I1001 16:52:03.081565 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"190b6fef6a0501c29d59acffb597e0d5cd54b2310aa997d6731fcc9f4534e864"} err="failed to get container status \"190b6fef6a0501c29d59acffb597e0d5cd54b2310aa997d6731fcc9f4534e864\": rpc error: code = NotFound desc = could not find container \"190b6fef6a0501c29d59acffb597e0d5cd54b2310aa997d6731fcc9f4534e864\": container with ID starting with 190b6fef6a0501c29d59acffb597e0d5cd54b2310aa997d6731fcc9f4534e864 not found: ID does not exist" Oct 01 16:52:03 crc kubenswrapper[4726]: I1001 16:52:03.824544 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1aead31-ac35-498f-b23f-cdfc25d72c88" path="/var/lib/kubelet/pods/d1aead31-ac35-498f-b23f-cdfc25d72c88/volumes" Oct 01 16:52:07 crc kubenswrapper[4726]: I1001 16:52:07.815585 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:52:07 crc kubenswrapper[4726]: E1001 16:52:07.816753 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:52:21 crc kubenswrapper[4726]: I1001 16:52:21.808662 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:52:21 crc kubenswrapper[4726]: E1001 16:52:21.809753 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:52:33 crc kubenswrapper[4726]: I1001 16:52:33.808640 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:52:33 crc kubenswrapper[4726]: E1001 16:52:33.809758 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:52:44 crc kubenswrapper[4726]: I1001 16:52:44.808692 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:52:44 crc kubenswrapper[4726]: E1001 16:52:44.809883 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:52:57 crc kubenswrapper[4726]: I1001 16:52:57.825625 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:52:57 crc kubenswrapper[4726]: E1001 16:52:57.826920 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:53:11 crc kubenswrapper[4726]: I1001 16:53:11.808523 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:53:11 crc kubenswrapper[4726]: E1001 16:53:11.809522 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:53:22 crc kubenswrapper[4726]: I1001 16:53:22.808037 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:53:22 crc kubenswrapper[4726]: E1001 16:53:22.808862 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:53:33 crc kubenswrapper[4726]: I1001 16:53:33.811624 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:53:34 crc kubenswrapper[4726]: I1001 16:53:34.942007 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"46f893977f5f67930718136271698b5b624cfc7f800e094b51bbecb8e5163132"} Oct 01 16:55:53 crc kubenswrapper[4726]: I1001 16:55:53.413363 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:55:53 crc kubenswrapper[4726]: I1001 16:55:53.413915 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:56:23 crc kubenswrapper[4726]: I1001 16:56:23.413362 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:56:23 crc kubenswrapper[4726]: I1001 16:56:23.413995 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:56:53 crc kubenswrapper[4726]: I1001 16:56:53.413591 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:56:53 crc kubenswrapper[4726]: I1001 16:56:53.414152 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:56:53 crc kubenswrapper[4726]: I1001 16:56:53.414200 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 16:56:53 crc kubenswrapper[4726]: I1001 16:56:53.414807 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"46f893977f5f67930718136271698b5b624cfc7f800e094b51bbecb8e5163132"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:56:53 crc kubenswrapper[4726]: I1001 16:56:53.414861 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://46f893977f5f67930718136271698b5b624cfc7f800e094b51bbecb8e5163132" gracePeriod=600 Oct 01 16:56:54 crc kubenswrapper[4726]: I1001 16:56:54.155470 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="46f893977f5f67930718136271698b5b624cfc7f800e094b51bbecb8e5163132" exitCode=0 Oct 01 16:56:54 crc kubenswrapper[4726]: I1001 16:56:54.155558 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"46f893977f5f67930718136271698b5b624cfc7f800e094b51bbecb8e5163132"} Oct 01 16:56:54 crc kubenswrapper[4726]: I1001 16:56:54.156096 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865"} Oct 01 16:56:54 crc kubenswrapper[4726]: I1001 16:56:54.156126 4726 scope.go:117] "RemoveContainer" containerID="0d19eb99e8b1459c077c55e4100a75bc010f90566d5adc2d5b61526a022c2636" Oct 01 16:56:58 crc kubenswrapper[4726]: I1001 16:56:58.919171 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5659n"] Oct 01 16:56:58 crc kubenswrapper[4726]: E1001 16:56:58.920359 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1aead31-ac35-498f-b23f-cdfc25d72c88" containerName="extract-utilities" Oct 01 16:56:58 crc kubenswrapper[4726]: I1001 16:56:58.920378 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1aead31-ac35-498f-b23f-cdfc25d72c88" containerName="extract-utilities" Oct 01 16:56:58 crc kubenswrapper[4726]: E1001 16:56:58.920391 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1aead31-ac35-498f-b23f-cdfc25d72c88" containerName="extract-content" Oct 01 16:56:58 crc kubenswrapper[4726]: I1001 16:56:58.920399 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1aead31-ac35-498f-b23f-cdfc25d72c88" containerName="extract-content" Oct 01 16:56:58 crc kubenswrapper[4726]: E1001 16:56:58.920434 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1aead31-ac35-498f-b23f-cdfc25d72c88" containerName="registry-server" Oct 01 16:56:58 crc kubenswrapper[4726]: I1001 16:56:58.920441 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1aead31-ac35-498f-b23f-cdfc25d72c88" containerName="registry-server" Oct 01 16:56:58 crc kubenswrapper[4726]: I1001 16:56:58.920664 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1aead31-ac35-498f-b23f-cdfc25d72c88" containerName="registry-server" Oct 01 16:56:58 crc kubenswrapper[4726]: I1001 16:56:58.922614 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:56:58 crc kubenswrapper[4726]: I1001 16:56:58.944471 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5659n"] Oct 01 16:56:59 crc kubenswrapper[4726]: I1001 16:56:59.063042 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9e363e0-8416-4717-b2fb-5767db0d8843-utilities\") pod \"certified-operators-5659n\" (UID: \"e9e363e0-8416-4717-b2fb-5767db0d8843\") " pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:56:59 crc kubenswrapper[4726]: I1001 16:56:59.063104 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czgj4\" (UniqueName: \"kubernetes.io/projected/e9e363e0-8416-4717-b2fb-5767db0d8843-kube-api-access-czgj4\") pod \"certified-operators-5659n\" (UID: \"e9e363e0-8416-4717-b2fb-5767db0d8843\") " pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:56:59 crc kubenswrapper[4726]: I1001 16:56:59.063187 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9e363e0-8416-4717-b2fb-5767db0d8843-catalog-content\") pod \"certified-operators-5659n\" (UID: \"e9e363e0-8416-4717-b2fb-5767db0d8843\") " pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:56:59 crc kubenswrapper[4726]: I1001 16:56:59.165316 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9e363e0-8416-4717-b2fb-5767db0d8843-utilities\") pod \"certified-operators-5659n\" (UID: \"e9e363e0-8416-4717-b2fb-5767db0d8843\") " pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:56:59 crc kubenswrapper[4726]: I1001 16:56:59.165376 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czgj4\" (UniqueName: \"kubernetes.io/projected/e9e363e0-8416-4717-b2fb-5767db0d8843-kube-api-access-czgj4\") pod \"certified-operators-5659n\" (UID: \"e9e363e0-8416-4717-b2fb-5767db0d8843\") " pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:56:59 crc kubenswrapper[4726]: I1001 16:56:59.165494 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9e363e0-8416-4717-b2fb-5767db0d8843-catalog-content\") pod \"certified-operators-5659n\" (UID: \"e9e363e0-8416-4717-b2fb-5767db0d8843\") " pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:56:59 crc kubenswrapper[4726]: I1001 16:56:59.165852 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9e363e0-8416-4717-b2fb-5767db0d8843-utilities\") pod \"certified-operators-5659n\" (UID: \"e9e363e0-8416-4717-b2fb-5767db0d8843\") " pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:56:59 crc kubenswrapper[4726]: I1001 16:56:59.166006 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9e363e0-8416-4717-b2fb-5767db0d8843-catalog-content\") pod \"certified-operators-5659n\" (UID: \"e9e363e0-8416-4717-b2fb-5767db0d8843\") " pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:56:59 crc kubenswrapper[4726]: I1001 16:56:59.191942 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czgj4\" (UniqueName: \"kubernetes.io/projected/e9e363e0-8416-4717-b2fb-5767db0d8843-kube-api-access-czgj4\") pod \"certified-operators-5659n\" (UID: \"e9e363e0-8416-4717-b2fb-5767db0d8843\") " pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:56:59 crc kubenswrapper[4726]: I1001 16:56:59.270066 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:56:59 crc kubenswrapper[4726]: I1001 16:56:59.794715 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5659n"] Oct 01 16:57:00 crc kubenswrapper[4726]: I1001 16:57:00.218572 4726 generic.go:334] "Generic (PLEG): container finished" podID="e9e363e0-8416-4717-b2fb-5767db0d8843" containerID="39ae339ac386a7ed0c1f6ce65204ff912007ba8b5cc0291ced8dedad156dff0f" exitCode=0 Oct 01 16:57:00 crc kubenswrapper[4726]: I1001 16:57:00.218648 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5659n" event={"ID":"e9e363e0-8416-4717-b2fb-5767db0d8843","Type":"ContainerDied","Data":"39ae339ac386a7ed0c1f6ce65204ff912007ba8b5cc0291ced8dedad156dff0f"} Oct 01 16:57:00 crc kubenswrapper[4726]: I1001 16:57:00.218910 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5659n" event={"ID":"e9e363e0-8416-4717-b2fb-5767db0d8843","Type":"ContainerStarted","Data":"9a8d3a23a11bdfa40a6d578f33183423be0ecf643c2cba8c9d528702ec61faaf"} Oct 01 16:57:00 crc kubenswrapper[4726]: I1001 16:57:00.221415 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:57:02 crc kubenswrapper[4726]: I1001 16:57:02.240014 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5659n" event={"ID":"e9e363e0-8416-4717-b2fb-5767db0d8843","Type":"ContainerStarted","Data":"0380490fb84b5df32fc4e1721ea30d7985d1347ead3d81f65c416b6c2dc98630"} Oct 01 16:57:08 crc kubenswrapper[4726]: I1001 16:57:08.307569 4726 generic.go:334] "Generic (PLEG): container finished" podID="e9e363e0-8416-4717-b2fb-5767db0d8843" containerID="0380490fb84b5df32fc4e1721ea30d7985d1347ead3d81f65c416b6c2dc98630" exitCode=0 Oct 01 16:57:08 crc kubenswrapper[4726]: I1001 16:57:08.307655 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5659n" event={"ID":"e9e363e0-8416-4717-b2fb-5767db0d8843","Type":"ContainerDied","Data":"0380490fb84b5df32fc4e1721ea30d7985d1347ead3d81f65c416b6c2dc98630"} Oct 01 16:57:09 crc kubenswrapper[4726]: I1001 16:57:09.322926 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5659n" event={"ID":"e9e363e0-8416-4717-b2fb-5767db0d8843","Type":"ContainerStarted","Data":"0689db711ab0fcbe310ccce6582b59b3523ca40dc1b1ccf8214e544a403a69f8"} Oct 01 16:57:09 crc kubenswrapper[4726]: I1001 16:57:09.342762 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5659n" podStartSLOduration=2.57679313 podStartE2EDuration="11.342741811s" podCreationTimestamp="2025-10-01 16:56:58 +0000 UTC" firstStartedPulling="2025-10-01 16:57:00.221129427 +0000 UTC m=+3473.122682004" lastFinishedPulling="2025-10-01 16:57:08.987078108 +0000 UTC m=+3481.888630685" observedRunningTime="2025-10-01 16:57:09.341208647 +0000 UTC m=+3482.242761234" watchObservedRunningTime="2025-10-01 16:57:09.342741811 +0000 UTC m=+3482.244294388" Oct 01 16:57:19 crc kubenswrapper[4726]: I1001 16:57:19.271161 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:57:19 crc kubenswrapper[4726]: I1001 16:57:19.271770 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:57:19 crc kubenswrapper[4726]: I1001 16:57:19.342581 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:57:19 crc kubenswrapper[4726]: I1001 16:57:19.476205 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:57:19 crc kubenswrapper[4726]: I1001 16:57:19.584504 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5659n"] Oct 01 16:57:21 crc kubenswrapper[4726]: I1001 16:57:21.441541 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5659n" podUID="e9e363e0-8416-4717-b2fb-5767db0d8843" containerName="registry-server" containerID="cri-o://0689db711ab0fcbe310ccce6582b59b3523ca40dc1b1ccf8214e544a403a69f8" gracePeriod=2 Oct 01 16:57:21 crc kubenswrapper[4726]: I1001 16:57:21.998103 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.169615 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9e363e0-8416-4717-b2fb-5767db0d8843-utilities\") pod \"e9e363e0-8416-4717-b2fb-5767db0d8843\" (UID: \"e9e363e0-8416-4717-b2fb-5767db0d8843\") " Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.169793 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9e363e0-8416-4717-b2fb-5767db0d8843-catalog-content\") pod \"e9e363e0-8416-4717-b2fb-5767db0d8843\" (UID: \"e9e363e0-8416-4717-b2fb-5767db0d8843\") " Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.169890 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czgj4\" (UniqueName: \"kubernetes.io/projected/e9e363e0-8416-4717-b2fb-5767db0d8843-kube-api-access-czgj4\") pod \"e9e363e0-8416-4717-b2fb-5767db0d8843\" (UID: \"e9e363e0-8416-4717-b2fb-5767db0d8843\") " Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.170832 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9e363e0-8416-4717-b2fb-5767db0d8843-utilities" (OuterVolumeSpecName: "utilities") pod "e9e363e0-8416-4717-b2fb-5767db0d8843" (UID: "e9e363e0-8416-4717-b2fb-5767db0d8843"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.179359 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9e363e0-8416-4717-b2fb-5767db0d8843-kube-api-access-czgj4" (OuterVolumeSpecName: "kube-api-access-czgj4") pod "e9e363e0-8416-4717-b2fb-5767db0d8843" (UID: "e9e363e0-8416-4717-b2fb-5767db0d8843"). InnerVolumeSpecName "kube-api-access-czgj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.243768 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9e363e0-8416-4717-b2fb-5767db0d8843-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9e363e0-8416-4717-b2fb-5767db0d8843" (UID: "e9e363e0-8416-4717-b2fb-5767db0d8843"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.271433 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9e363e0-8416-4717-b2fb-5767db0d8843-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.271460 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czgj4\" (UniqueName: \"kubernetes.io/projected/e9e363e0-8416-4717-b2fb-5767db0d8843-kube-api-access-czgj4\") on node \"crc\" DevicePath \"\"" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.271471 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9e363e0-8416-4717-b2fb-5767db0d8843-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.455096 4726 generic.go:334] "Generic (PLEG): container finished" podID="e9e363e0-8416-4717-b2fb-5767db0d8843" containerID="0689db711ab0fcbe310ccce6582b59b3523ca40dc1b1ccf8214e544a403a69f8" exitCode=0 Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.455141 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5659n" event={"ID":"e9e363e0-8416-4717-b2fb-5767db0d8843","Type":"ContainerDied","Data":"0689db711ab0fcbe310ccce6582b59b3523ca40dc1b1ccf8214e544a403a69f8"} Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.455173 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5659n" event={"ID":"e9e363e0-8416-4717-b2fb-5767db0d8843","Type":"ContainerDied","Data":"9a8d3a23a11bdfa40a6d578f33183423be0ecf643c2cba8c9d528702ec61faaf"} Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.455195 4726 scope.go:117] "RemoveContainer" containerID="0689db711ab0fcbe310ccce6582b59b3523ca40dc1b1ccf8214e544a403a69f8" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.455192 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5659n" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.511423 4726 scope.go:117] "RemoveContainer" containerID="0380490fb84b5df32fc4e1721ea30d7985d1347ead3d81f65c416b6c2dc98630" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.522810 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5659n"] Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.540414 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5659n"] Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.540912 4726 scope.go:117] "RemoveContainer" containerID="39ae339ac386a7ed0c1f6ce65204ff912007ba8b5cc0291ced8dedad156dff0f" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.578576 4726 scope.go:117] "RemoveContainer" containerID="0689db711ab0fcbe310ccce6582b59b3523ca40dc1b1ccf8214e544a403a69f8" Oct 01 16:57:22 crc kubenswrapper[4726]: E1001 16:57:22.587229 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0689db711ab0fcbe310ccce6582b59b3523ca40dc1b1ccf8214e544a403a69f8\": container with ID starting with 0689db711ab0fcbe310ccce6582b59b3523ca40dc1b1ccf8214e544a403a69f8 not found: ID does not exist" containerID="0689db711ab0fcbe310ccce6582b59b3523ca40dc1b1ccf8214e544a403a69f8" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.587291 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0689db711ab0fcbe310ccce6582b59b3523ca40dc1b1ccf8214e544a403a69f8"} err="failed to get container status \"0689db711ab0fcbe310ccce6582b59b3523ca40dc1b1ccf8214e544a403a69f8\": rpc error: code = NotFound desc = could not find container \"0689db711ab0fcbe310ccce6582b59b3523ca40dc1b1ccf8214e544a403a69f8\": container with ID starting with 0689db711ab0fcbe310ccce6582b59b3523ca40dc1b1ccf8214e544a403a69f8 not found: ID does not exist" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.587326 4726 scope.go:117] "RemoveContainer" containerID="0380490fb84b5df32fc4e1721ea30d7985d1347ead3d81f65c416b6c2dc98630" Oct 01 16:57:22 crc kubenswrapper[4726]: E1001 16:57:22.587672 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0380490fb84b5df32fc4e1721ea30d7985d1347ead3d81f65c416b6c2dc98630\": container with ID starting with 0380490fb84b5df32fc4e1721ea30d7985d1347ead3d81f65c416b6c2dc98630 not found: ID does not exist" containerID="0380490fb84b5df32fc4e1721ea30d7985d1347ead3d81f65c416b6c2dc98630" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.587695 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0380490fb84b5df32fc4e1721ea30d7985d1347ead3d81f65c416b6c2dc98630"} err="failed to get container status \"0380490fb84b5df32fc4e1721ea30d7985d1347ead3d81f65c416b6c2dc98630\": rpc error: code = NotFound desc = could not find container \"0380490fb84b5df32fc4e1721ea30d7985d1347ead3d81f65c416b6c2dc98630\": container with ID starting with 0380490fb84b5df32fc4e1721ea30d7985d1347ead3d81f65c416b6c2dc98630 not found: ID does not exist" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.587714 4726 scope.go:117] "RemoveContainer" containerID="39ae339ac386a7ed0c1f6ce65204ff912007ba8b5cc0291ced8dedad156dff0f" Oct 01 16:57:22 crc kubenswrapper[4726]: E1001 16:57:22.588002 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39ae339ac386a7ed0c1f6ce65204ff912007ba8b5cc0291ced8dedad156dff0f\": container with ID starting with 39ae339ac386a7ed0c1f6ce65204ff912007ba8b5cc0291ced8dedad156dff0f not found: ID does not exist" containerID="39ae339ac386a7ed0c1f6ce65204ff912007ba8b5cc0291ced8dedad156dff0f" Oct 01 16:57:22 crc kubenswrapper[4726]: I1001 16:57:22.588022 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39ae339ac386a7ed0c1f6ce65204ff912007ba8b5cc0291ced8dedad156dff0f"} err="failed to get container status \"39ae339ac386a7ed0c1f6ce65204ff912007ba8b5cc0291ced8dedad156dff0f\": rpc error: code = NotFound desc = could not find container \"39ae339ac386a7ed0c1f6ce65204ff912007ba8b5cc0291ced8dedad156dff0f\": container with ID starting with 39ae339ac386a7ed0c1f6ce65204ff912007ba8b5cc0291ced8dedad156dff0f not found: ID does not exist" Oct 01 16:57:23 crc kubenswrapper[4726]: I1001 16:57:23.820817 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9e363e0-8416-4717-b2fb-5767db0d8843" path="/var/lib/kubelet/pods/e9e363e0-8416-4717-b2fb-5767db0d8843/volumes" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.035416 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gkxct"] Oct 01 16:58:15 crc kubenswrapper[4726]: E1001 16:58:15.036673 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9e363e0-8416-4717-b2fb-5767db0d8843" containerName="registry-server" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.036697 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9e363e0-8416-4717-b2fb-5767db0d8843" containerName="registry-server" Oct 01 16:58:15 crc kubenswrapper[4726]: E1001 16:58:15.036722 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9e363e0-8416-4717-b2fb-5767db0d8843" containerName="extract-utilities" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.036733 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9e363e0-8416-4717-b2fb-5767db0d8843" containerName="extract-utilities" Oct 01 16:58:15 crc kubenswrapper[4726]: E1001 16:58:15.036771 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9e363e0-8416-4717-b2fb-5767db0d8843" containerName="extract-content" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.036782 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9e363e0-8416-4717-b2fb-5767db0d8843" containerName="extract-content" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.037140 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9e363e0-8416-4717-b2fb-5767db0d8843" containerName="registry-server" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.039432 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.056548 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkxct"] Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.120916 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0bac0b4-1db2-4f8f-a226-771135977a8b-utilities\") pod \"redhat-marketplace-gkxct\" (UID: \"a0bac0b4-1db2-4f8f-a226-771135977a8b\") " pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.121405 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfs2t\" (UniqueName: \"kubernetes.io/projected/a0bac0b4-1db2-4f8f-a226-771135977a8b-kube-api-access-mfs2t\") pod \"redhat-marketplace-gkxct\" (UID: \"a0bac0b4-1db2-4f8f-a226-771135977a8b\") " pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.121674 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0bac0b4-1db2-4f8f-a226-771135977a8b-catalog-content\") pod \"redhat-marketplace-gkxct\" (UID: \"a0bac0b4-1db2-4f8f-a226-771135977a8b\") " pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.223002 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4224x"] Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.223763 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfs2t\" (UniqueName: \"kubernetes.io/projected/a0bac0b4-1db2-4f8f-a226-771135977a8b-kube-api-access-mfs2t\") pod \"redhat-marketplace-gkxct\" (UID: \"a0bac0b4-1db2-4f8f-a226-771135977a8b\") " pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.223817 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0bac0b4-1db2-4f8f-a226-771135977a8b-catalog-content\") pod \"redhat-marketplace-gkxct\" (UID: \"a0bac0b4-1db2-4f8f-a226-771135977a8b\") " pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.223849 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0bac0b4-1db2-4f8f-a226-771135977a8b-utilities\") pod \"redhat-marketplace-gkxct\" (UID: \"a0bac0b4-1db2-4f8f-a226-771135977a8b\") " pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.224454 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0bac0b4-1db2-4f8f-a226-771135977a8b-catalog-content\") pod \"redhat-marketplace-gkxct\" (UID: \"a0bac0b4-1db2-4f8f-a226-771135977a8b\") " pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.224479 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0bac0b4-1db2-4f8f-a226-771135977a8b-utilities\") pod \"redhat-marketplace-gkxct\" (UID: \"a0bac0b4-1db2-4f8f-a226-771135977a8b\") " pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.225454 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.240916 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4224x"] Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.251218 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfs2t\" (UniqueName: \"kubernetes.io/projected/a0bac0b4-1db2-4f8f-a226-771135977a8b-kube-api-access-mfs2t\") pod \"redhat-marketplace-gkxct\" (UID: \"a0bac0b4-1db2-4f8f-a226-771135977a8b\") " pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.325472 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9cxc\" (UniqueName: \"kubernetes.io/projected/54e641e4-4402-4da7-aa1c-159714850518-kube-api-access-t9cxc\") pod \"community-operators-4224x\" (UID: \"54e641e4-4402-4da7-aa1c-159714850518\") " pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.325808 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54e641e4-4402-4da7-aa1c-159714850518-utilities\") pod \"community-operators-4224x\" (UID: \"54e641e4-4402-4da7-aa1c-159714850518\") " pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.325980 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54e641e4-4402-4da7-aa1c-159714850518-catalog-content\") pod \"community-operators-4224x\" (UID: \"54e641e4-4402-4da7-aa1c-159714850518\") " pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.390618 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.428753 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54e641e4-4402-4da7-aa1c-159714850518-catalog-content\") pod \"community-operators-4224x\" (UID: \"54e641e4-4402-4da7-aa1c-159714850518\") " pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.429705 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54e641e4-4402-4da7-aa1c-159714850518-catalog-content\") pod \"community-operators-4224x\" (UID: \"54e641e4-4402-4da7-aa1c-159714850518\") " pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.430394 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9cxc\" (UniqueName: \"kubernetes.io/projected/54e641e4-4402-4da7-aa1c-159714850518-kube-api-access-t9cxc\") pod \"community-operators-4224x\" (UID: \"54e641e4-4402-4da7-aa1c-159714850518\") " pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.430913 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54e641e4-4402-4da7-aa1c-159714850518-utilities\") pod \"community-operators-4224x\" (UID: \"54e641e4-4402-4da7-aa1c-159714850518\") " pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.431310 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54e641e4-4402-4da7-aa1c-159714850518-utilities\") pod \"community-operators-4224x\" (UID: \"54e641e4-4402-4da7-aa1c-159714850518\") " pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.459095 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9cxc\" (UniqueName: \"kubernetes.io/projected/54e641e4-4402-4da7-aa1c-159714850518-kube-api-access-t9cxc\") pod \"community-operators-4224x\" (UID: \"54e641e4-4402-4da7-aa1c-159714850518\") " pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.551461 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:15 crc kubenswrapper[4726]: I1001 16:58:15.892549 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkxct"] Oct 01 16:58:16 crc kubenswrapper[4726]: I1001 16:58:16.048418 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkxct" event={"ID":"a0bac0b4-1db2-4f8f-a226-771135977a8b","Type":"ContainerStarted","Data":"6fba43d962003fcff90748f09151a644e8fc90c7a49d6fd0c3bc112f892c2a78"} Oct 01 16:58:16 crc kubenswrapper[4726]: I1001 16:58:16.075405 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4224x"] Oct 01 16:58:16 crc kubenswrapper[4726]: W1001 16:58:16.138081 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54e641e4_4402_4da7_aa1c_159714850518.slice/crio-dd11e010a324ea500612f4fe3ceaba93b3ff50d8e9088a92441487e9ce609817 WatchSource:0}: Error finding container dd11e010a324ea500612f4fe3ceaba93b3ff50d8e9088a92441487e9ce609817: Status 404 returned error can't find the container with id dd11e010a324ea500612f4fe3ceaba93b3ff50d8e9088a92441487e9ce609817 Oct 01 16:58:17 crc kubenswrapper[4726]: I1001 16:58:17.058706 4726 generic.go:334] "Generic (PLEG): container finished" podID="54e641e4-4402-4da7-aa1c-159714850518" containerID="7e9fcfbedb5f4082e450c92c7a7283b5ec625c230f9d30a5f2d2a073f7c4bd2e" exitCode=0 Oct 01 16:58:17 crc kubenswrapper[4726]: I1001 16:58:17.058779 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4224x" event={"ID":"54e641e4-4402-4da7-aa1c-159714850518","Type":"ContainerDied","Data":"7e9fcfbedb5f4082e450c92c7a7283b5ec625c230f9d30a5f2d2a073f7c4bd2e"} Oct 01 16:58:17 crc kubenswrapper[4726]: I1001 16:58:17.059072 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4224x" event={"ID":"54e641e4-4402-4da7-aa1c-159714850518","Type":"ContainerStarted","Data":"dd11e010a324ea500612f4fe3ceaba93b3ff50d8e9088a92441487e9ce609817"} Oct 01 16:58:17 crc kubenswrapper[4726]: I1001 16:58:17.061001 4726 generic.go:334] "Generic (PLEG): container finished" podID="a0bac0b4-1db2-4f8f-a226-771135977a8b" containerID="5ce0483a11225894ab83e525cbe90af6f7cbc5c02826a165ad9a5adfbff3e4e6" exitCode=0 Oct 01 16:58:17 crc kubenswrapper[4726]: I1001 16:58:17.061041 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkxct" event={"ID":"a0bac0b4-1db2-4f8f-a226-771135977a8b","Type":"ContainerDied","Data":"5ce0483a11225894ab83e525cbe90af6f7cbc5c02826a165ad9a5adfbff3e4e6"} Oct 01 16:58:18 crc kubenswrapper[4726]: E1001 16:58:18.785946 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54e641e4_4402_4da7_aa1c_159714850518.slice/crio-conmon-9a47cf0e9523a6653573f7c92627ac3ab724f143c956e616ef3afe694a6219c8.scope\": RecentStats: unable to find data in memory cache]" Oct 01 16:58:19 crc kubenswrapper[4726]: I1001 16:58:19.080518 4726 generic.go:334] "Generic (PLEG): container finished" podID="a0bac0b4-1db2-4f8f-a226-771135977a8b" containerID="95014e024f548b5fa727d478e775e312934606f81c9caa287b0ff0820a64526c" exitCode=0 Oct 01 16:58:19 crc kubenswrapper[4726]: I1001 16:58:19.080556 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkxct" event={"ID":"a0bac0b4-1db2-4f8f-a226-771135977a8b","Type":"ContainerDied","Data":"95014e024f548b5fa727d478e775e312934606f81c9caa287b0ff0820a64526c"} Oct 01 16:58:19 crc kubenswrapper[4726]: I1001 16:58:19.084067 4726 generic.go:334] "Generic (PLEG): container finished" podID="54e641e4-4402-4da7-aa1c-159714850518" containerID="9a47cf0e9523a6653573f7c92627ac3ab724f143c956e616ef3afe694a6219c8" exitCode=0 Oct 01 16:58:19 crc kubenswrapper[4726]: I1001 16:58:19.084102 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4224x" event={"ID":"54e641e4-4402-4da7-aa1c-159714850518","Type":"ContainerDied","Data":"9a47cf0e9523a6653573f7c92627ac3ab724f143c956e616ef3afe694a6219c8"} Oct 01 16:58:20 crc kubenswrapper[4726]: I1001 16:58:20.094333 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkxct" event={"ID":"a0bac0b4-1db2-4f8f-a226-771135977a8b","Type":"ContainerStarted","Data":"7885937c6dd7d8bccb780ddd80e60a7c039174ea3dd3a5f68238c9d7baf9be84"} Oct 01 16:58:20 crc kubenswrapper[4726]: I1001 16:58:20.099419 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4224x" event={"ID":"54e641e4-4402-4da7-aa1c-159714850518","Type":"ContainerStarted","Data":"2dfe8c54fec32f3c126adc748a3d6c96b24c27de93fbe26e5a1988db8b3d4ae6"} Oct 01 16:58:20 crc kubenswrapper[4726]: I1001 16:58:20.114081 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gkxct" podStartSLOduration=2.502432039 podStartE2EDuration="5.11406187s" podCreationTimestamp="2025-10-01 16:58:15 +0000 UTC" firstStartedPulling="2025-10-01 16:58:17.063034572 +0000 UTC m=+3549.964587149" lastFinishedPulling="2025-10-01 16:58:19.674664393 +0000 UTC m=+3552.576216980" observedRunningTime="2025-10-01 16:58:20.110347473 +0000 UTC m=+3553.011900060" watchObservedRunningTime="2025-10-01 16:58:20.11406187 +0000 UTC m=+3553.015614447" Oct 01 16:58:20 crc kubenswrapper[4726]: I1001 16:58:20.131067 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4224x" podStartSLOduration=2.669014212 podStartE2EDuration="5.131027559s" podCreationTimestamp="2025-10-01 16:58:15 +0000 UTC" firstStartedPulling="2025-10-01 16:58:17.060913621 +0000 UTC m=+3549.962466198" lastFinishedPulling="2025-10-01 16:58:19.522926958 +0000 UTC m=+3552.424479545" observedRunningTime="2025-10-01 16:58:20.130343419 +0000 UTC m=+3553.031896036" watchObservedRunningTime="2025-10-01 16:58:20.131027559 +0000 UTC m=+3553.032580136" Oct 01 16:58:25 crc kubenswrapper[4726]: I1001 16:58:25.391038 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:25 crc kubenswrapper[4726]: I1001 16:58:25.394244 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:25 crc kubenswrapper[4726]: I1001 16:58:25.466630 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:25 crc kubenswrapper[4726]: I1001 16:58:25.553522 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:25 crc kubenswrapper[4726]: I1001 16:58:25.553594 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:25 crc kubenswrapper[4726]: I1001 16:58:25.627932 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:26 crc kubenswrapper[4726]: I1001 16:58:26.273965 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:26 crc kubenswrapper[4726]: I1001 16:58:26.278982 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:26 crc kubenswrapper[4726]: I1001 16:58:26.922654 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4224x"] Oct 01 16:58:28 crc kubenswrapper[4726]: I1001 16:58:28.217648 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4224x" podUID="54e641e4-4402-4da7-aa1c-159714850518" containerName="registry-server" containerID="cri-o://2dfe8c54fec32f3c126adc748a3d6c96b24c27de93fbe26e5a1988db8b3d4ae6" gracePeriod=2 Oct 01 16:58:28 crc kubenswrapper[4726]: I1001 16:58:28.719826 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkxct"] Oct 01 16:58:28 crc kubenswrapper[4726]: I1001 16:58:28.761831 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:28 crc kubenswrapper[4726]: I1001 16:58:28.860276 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54e641e4-4402-4da7-aa1c-159714850518-utilities\") pod \"54e641e4-4402-4da7-aa1c-159714850518\" (UID: \"54e641e4-4402-4da7-aa1c-159714850518\") " Oct 01 16:58:28 crc kubenswrapper[4726]: I1001 16:58:28.860473 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9cxc\" (UniqueName: \"kubernetes.io/projected/54e641e4-4402-4da7-aa1c-159714850518-kube-api-access-t9cxc\") pod \"54e641e4-4402-4da7-aa1c-159714850518\" (UID: \"54e641e4-4402-4da7-aa1c-159714850518\") " Oct 01 16:58:28 crc kubenswrapper[4726]: I1001 16:58:28.860674 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54e641e4-4402-4da7-aa1c-159714850518-catalog-content\") pod \"54e641e4-4402-4da7-aa1c-159714850518\" (UID: \"54e641e4-4402-4da7-aa1c-159714850518\") " Oct 01 16:58:28 crc kubenswrapper[4726]: I1001 16:58:28.862907 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54e641e4-4402-4da7-aa1c-159714850518-utilities" (OuterVolumeSpecName: "utilities") pod "54e641e4-4402-4da7-aa1c-159714850518" (UID: "54e641e4-4402-4da7-aa1c-159714850518"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:58:28 crc kubenswrapper[4726]: I1001 16:58:28.868544 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54e641e4-4402-4da7-aa1c-159714850518-kube-api-access-t9cxc" (OuterVolumeSpecName: "kube-api-access-t9cxc") pod "54e641e4-4402-4da7-aa1c-159714850518" (UID: "54e641e4-4402-4da7-aa1c-159714850518"). InnerVolumeSpecName "kube-api-access-t9cxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:58:28 crc kubenswrapper[4726]: I1001 16:58:28.963702 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9cxc\" (UniqueName: \"kubernetes.io/projected/54e641e4-4402-4da7-aa1c-159714850518-kube-api-access-t9cxc\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:28 crc kubenswrapper[4726]: I1001 16:58:28.963738 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54e641e4-4402-4da7-aa1c-159714850518-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.235674 4726 generic.go:334] "Generic (PLEG): container finished" podID="54e641e4-4402-4da7-aa1c-159714850518" containerID="2dfe8c54fec32f3c126adc748a3d6c96b24c27de93fbe26e5a1988db8b3d4ae6" exitCode=0 Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.237295 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gkxct" podUID="a0bac0b4-1db2-4f8f-a226-771135977a8b" containerName="registry-server" containerID="cri-o://7885937c6dd7d8bccb780ddd80e60a7c039174ea3dd3a5f68238c9d7baf9be84" gracePeriod=2 Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.235812 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4224x" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.235850 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4224x" event={"ID":"54e641e4-4402-4da7-aa1c-159714850518","Type":"ContainerDied","Data":"2dfe8c54fec32f3c126adc748a3d6c96b24c27de93fbe26e5a1988db8b3d4ae6"} Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.239205 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4224x" event={"ID":"54e641e4-4402-4da7-aa1c-159714850518","Type":"ContainerDied","Data":"dd11e010a324ea500612f4fe3ceaba93b3ff50d8e9088a92441487e9ce609817"} Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.239247 4726 scope.go:117] "RemoveContainer" containerID="2dfe8c54fec32f3c126adc748a3d6c96b24c27de93fbe26e5a1988db8b3d4ae6" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.277773 4726 scope.go:117] "RemoveContainer" containerID="9a47cf0e9523a6653573f7c92627ac3ab724f143c956e616ef3afe694a6219c8" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.319912 4726 scope.go:117] "RemoveContainer" containerID="7e9fcfbedb5f4082e450c92c7a7283b5ec625c230f9d30a5f2d2a073f7c4bd2e" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.366651 4726 scope.go:117] "RemoveContainer" containerID="2dfe8c54fec32f3c126adc748a3d6c96b24c27de93fbe26e5a1988db8b3d4ae6" Oct 01 16:58:29 crc kubenswrapper[4726]: E1001 16:58:29.367443 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dfe8c54fec32f3c126adc748a3d6c96b24c27de93fbe26e5a1988db8b3d4ae6\": container with ID starting with 2dfe8c54fec32f3c126adc748a3d6c96b24c27de93fbe26e5a1988db8b3d4ae6 not found: ID does not exist" containerID="2dfe8c54fec32f3c126adc748a3d6c96b24c27de93fbe26e5a1988db8b3d4ae6" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.367496 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dfe8c54fec32f3c126adc748a3d6c96b24c27de93fbe26e5a1988db8b3d4ae6"} err="failed to get container status \"2dfe8c54fec32f3c126adc748a3d6c96b24c27de93fbe26e5a1988db8b3d4ae6\": rpc error: code = NotFound desc = could not find container \"2dfe8c54fec32f3c126adc748a3d6c96b24c27de93fbe26e5a1988db8b3d4ae6\": container with ID starting with 2dfe8c54fec32f3c126adc748a3d6c96b24c27de93fbe26e5a1988db8b3d4ae6 not found: ID does not exist" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.367527 4726 scope.go:117] "RemoveContainer" containerID="9a47cf0e9523a6653573f7c92627ac3ab724f143c956e616ef3afe694a6219c8" Oct 01 16:58:29 crc kubenswrapper[4726]: E1001 16:58:29.368012 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a47cf0e9523a6653573f7c92627ac3ab724f143c956e616ef3afe694a6219c8\": container with ID starting with 9a47cf0e9523a6653573f7c92627ac3ab724f143c956e616ef3afe694a6219c8 not found: ID does not exist" containerID="9a47cf0e9523a6653573f7c92627ac3ab724f143c956e616ef3afe694a6219c8" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.368060 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a47cf0e9523a6653573f7c92627ac3ab724f143c956e616ef3afe694a6219c8"} err="failed to get container status \"9a47cf0e9523a6653573f7c92627ac3ab724f143c956e616ef3afe694a6219c8\": rpc error: code = NotFound desc = could not find container \"9a47cf0e9523a6653573f7c92627ac3ab724f143c956e616ef3afe694a6219c8\": container with ID starting with 9a47cf0e9523a6653573f7c92627ac3ab724f143c956e616ef3afe694a6219c8 not found: ID does not exist" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.368083 4726 scope.go:117] "RemoveContainer" containerID="7e9fcfbedb5f4082e450c92c7a7283b5ec625c230f9d30a5f2d2a073f7c4bd2e" Oct 01 16:58:29 crc kubenswrapper[4726]: E1001 16:58:29.368448 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e9fcfbedb5f4082e450c92c7a7283b5ec625c230f9d30a5f2d2a073f7c4bd2e\": container with ID starting with 7e9fcfbedb5f4082e450c92c7a7283b5ec625c230f9d30a5f2d2a073f7c4bd2e not found: ID does not exist" containerID="7e9fcfbedb5f4082e450c92c7a7283b5ec625c230f9d30a5f2d2a073f7c4bd2e" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.368479 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e9fcfbedb5f4082e450c92c7a7283b5ec625c230f9d30a5f2d2a073f7c4bd2e"} err="failed to get container status \"7e9fcfbedb5f4082e450c92c7a7283b5ec625c230f9d30a5f2d2a073f7c4bd2e\": rpc error: code = NotFound desc = could not find container \"7e9fcfbedb5f4082e450c92c7a7283b5ec625c230f9d30a5f2d2a073f7c4bd2e\": container with ID starting with 7e9fcfbedb5f4082e450c92c7a7283b5ec625c230f9d30a5f2d2a073f7c4bd2e not found: ID does not exist" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.390297 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54e641e4-4402-4da7-aa1c-159714850518-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54e641e4-4402-4da7-aa1c-159714850518" (UID: "54e641e4-4402-4da7-aa1c-159714850518"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.485264 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54e641e4-4402-4da7-aa1c-159714850518-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.615594 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4224x"] Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.627150 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4224x"] Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.817883 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54e641e4-4402-4da7-aa1c-159714850518" path="/var/lib/kubelet/pods/54e641e4-4402-4da7-aa1c-159714850518/volumes" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.870712 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.994883 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfs2t\" (UniqueName: \"kubernetes.io/projected/a0bac0b4-1db2-4f8f-a226-771135977a8b-kube-api-access-mfs2t\") pod \"a0bac0b4-1db2-4f8f-a226-771135977a8b\" (UID: \"a0bac0b4-1db2-4f8f-a226-771135977a8b\") " Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.995243 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0bac0b4-1db2-4f8f-a226-771135977a8b-catalog-content\") pod \"a0bac0b4-1db2-4f8f-a226-771135977a8b\" (UID: \"a0bac0b4-1db2-4f8f-a226-771135977a8b\") " Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.995272 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0bac0b4-1db2-4f8f-a226-771135977a8b-utilities\") pod \"a0bac0b4-1db2-4f8f-a226-771135977a8b\" (UID: \"a0bac0b4-1db2-4f8f-a226-771135977a8b\") " Oct 01 16:58:29 crc kubenswrapper[4726]: I1001 16:58:29.996401 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0bac0b4-1db2-4f8f-a226-771135977a8b-utilities" (OuterVolumeSpecName: "utilities") pod "a0bac0b4-1db2-4f8f-a226-771135977a8b" (UID: "a0bac0b4-1db2-4f8f-a226-771135977a8b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.002335 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0bac0b4-1db2-4f8f-a226-771135977a8b-kube-api-access-mfs2t" (OuterVolumeSpecName: "kube-api-access-mfs2t") pod "a0bac0b4-1db2-4f8f-a226-771135977a8b" (UID: "a0bac0b4-1db2-4f8f-a226-771135977a8b"). InnerVolumeSpecName "kube-api-access-mfs2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.010792 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0bac0b4-1db2-4f8f-a226-771135977a8b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0bac0b4-1db2-4f8f-a226-771135977a8b" (UID: "a0bac0b4-1db2-4f8f-a226-771135977a8b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.097313 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfs2t\" (UniqueName: \"kubernetes.io/projected/a0bac0b4-1db2-4f8f-a226-771135977a8b-kube-api-access-mfs2t\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.097353 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0bac0b4-1db2-4f8f-a226-771135977a8b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.097363 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0bac0b4-1db2-4f8f-a226-771135977a8b-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.258659 4726 generic.go:334] "Generic (PLEG): container finished" podID="a0bac0b4-1db2-4f8f-a226-771135977a8b" containerID="7885937c6dd7d8bccb780ddd80e60a7c039174ea3dd3a5f68238c9d7baf9be84" exitCode=0 Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.258724 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkxct" event={"ID":"a0bac0b4-1db2-4f8f-a226-771135977a8b","Type":"ContainerDied","Data":"7885937c6dd7d8bccb780ddd80e60a7c039174ea3dd3a5f68238c9d7baf9be84"} Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.258762 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gkxct" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.258809 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkxct" event={"ID":"a0bac0b4-1db2-4f8f-a226-771135977a8b","Type":"ContainerDied","Data":"6fba43d962003fcff90748f09151a644e8fc90c7a49d6fd0c3bc112f892c2a78"} Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.259037 4726 scope.go:117] "RemoveContainer" containerID="7885937c6dd7d8bccb780ddd80e60a7c039174ea3dd3a5f68238c9d7baf9be84" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.295671 4726 scope.go:117] "RemoveContainer" containerID="95014e024f548b5fa727d478e775e312934606f81c9caa287b0ff0820a64526c" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.335615 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkxct"] Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.350316 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkxct"] Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.366290 4726 scope.go:117] "RemoveContainer" containerID="5ce0483a11225894ab83e525cbe90af6f7cbc5c02826a165ad9a5adfbff3e4e6" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.424646 4726 scope.go:117] "RemoveContainer" containerID="7885937c6dd7d8bccb780ddd80e60a7c039174ea3dd3a5f68238c9d7baf9be84" Oct 01 16:58:30 crc kubenswrapper[4726]: E1001 16:58:30.425473 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7885937c6dd7d8bccb780ddd80e60a7c039174ea3dd3a5f68238c9d7baf9be84\": container with ID starting with 7885937c6dd7d8bccb780ddd80e60a7c039174ea3dd3a5f68238c9d7baf9be84 not found: ID does not exist" containerID="7885937c6dd7d8bccb780ddd80e60a7c039174ea3dd3a5f68238c9d7baf9be84" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.425543 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7885937c6dd7d8bccb780ddd80e60a7c039174ea3dd3a5f68238c9d7baf9be84"} err="failed to get container status \"7885937c6dd7d8bccb780ddd80e60a7c039174ea3dd3a5f68238c9d7baf9be84\": rpc error: code = NotFound desc = could not find container \"7885937c6dd7d8bccb780ddd80e60a7c039174ea3dd3a5f68238c9d7baf9be84\": container with ID starting with 7885937c6dd7d8bccb780ddd80e60a7c039174ea3dd3a5f68238c9d7baf9be84 not found: ID does not exist" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.425587 4726 scope.go:117] "RemoveContainer" containerID="95014e024f548b5fa727d478e775e312934606f81c9caa287b0ff0820a64526c" Oct 01 16:58:30 crc kubenswrapper[4726]: E1001 16:58:30.426070 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95014e024f548b5fa727d478e775e312934606f81c9caa287b0ff0820a64526c\": container with ID starting with 95014e024f548b5fa727d478e775e312934606f81c9caa287b0ff0820a64526c not found: ID does not exist" containerID="95014e024f548b5fa727d478e775e312934606f81c9caa287b0ff0820a64526c" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.426119 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95014e024f548b5fa727d478e775e312934606f81c9caa287b0ff0820a64526c"} err="failed to get container status \"95014e024f548b5fa727d478e775e312934606f81c9caa287b0ff0820a64526c\": rpc error: code = NotFound desc = could not find container \"95014e024f548b5fa727d478e775e312934606f81c9caa287b0ff0820a64526c\": container with ID starting with 95014e024f548b5fa727d478e775e312934606f81c9caa287b0ff0820a64526c not found: ID does not exist" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.426202 4726 scope.go:117] "RemoveContainer" containerID="5ce0483a11225894ab83e525cbe90af6f7cbc5c02826a165ad9a5adfbff3e4e6" Oct 01 16:58:30 crc kubenswrapper[4726]: E1001 16:58:30.426534 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ce0483a11225894ab83e525cbe90af6f7cbc5c02826a165ad9a5adfbff3e4e6\": container with ID starting with 5ce0483a11225894ab83e525cbe90af6f7cbc5c02826a165ad9a5adfbff3e4e6 not found: ID does not exist" containerID="5ce0483a11225894ab83e525cbe90af6f7cbc5c02826a165ad9a5adfbff3e4e6" Oct 01 16:58:30 crc kubenswrapper[4726]: I1001 16:58:30.426585 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ce0483a11225894ab83e525cbe90af6f7cbc5c02826a165ad9a5adfbff3e4e6"} err="failed to get container status \"5ce0483a11225894ab83e525cbe90af6f7cbc5c02826a165ad9a5adfbff3e4e6\": rpc error: code = NotFound desc = could not find container \"5ce0483a11225894ab83e525cbe90af6f7cbc5c02826a165ad9a5adfbff3e4e6\": container with ID starting with 5ce0483a11225894ab83e525cbe90af6f7cbc5c02826a165ad9a5adfbff3e4e6 not found: ID does not exist" Oct 01 16:58:31 crc kubenswrapper[4726]: I1001 16:58:31.817991 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0bac0b4-1db2-4f8f-a226-771135977a8b" path="/var/lib/kubelet/pods/a0bac0b4-1db2-4f8f-a226-771135977a8b/volumes" Oct 01 16:58:45 crc kubenswrapper[4726]: I1001 16:58:45.422824 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"261dc564-2126-415f-a5ce-a988af8a053c","Type":"ContainerDied","Data":"2b9050298664a40b33ab5e352d7bec523157f3dd66ab03da1a7a0c5e208f08ed"} Oct 01 16:58:45 crc kubenswrapper[4726]: I1001 16:58:45.422762 4726 generic.go:334] "Generic (PLEG): container finished" podID="261dc564-2126-415f-a5ce-a988af8a053c" containerID="2b9050298664a40b33ab5e352d7bec523157f3dd66ab03da1a7a0c5e208f08ed" exitCode=0 Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.903577 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.961546 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-openstack-config-secret\") pod \"261dc564-2126-415f-a5ce-a988af8a053c\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.961840 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/261dc564-2126-415f-a5ce-a988af8a053c-config-data\") pod \"261dc564-2126-415f-a5ce-a988af8a053c\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.961968 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-ssh-key\") pod \"261dc564-2126-415f-a5ce-a988af8a053c\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.962162 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-ca-certs\") pod \"261dc564-2126-415f-a5ce-a988af8a053c\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.962306 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmlxv\" (UniqueName: \"kubernetes.io/projected/261dc564-2126-415f-a5ce-a988af8a053c-kube-api-access-tmlxv\") pod \"261dc564-2126-415f-a5ce-a988af8a053c\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.962407 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/261dc564-2126-415f-a5ce-a988af8a053c-openstack-config\") pod \"261dc564-2126-415f-a5ce-a988af8a053c\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.962507 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"261dc564-2126-415f-a5ce-a988af8a053c\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.962652 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/261dc564-2126-415f-a5ce-a988af8a053c-test-operator-ephemeral-temporary\") pod \"261dc564-2126-415f-a5ce-a988af8a053c\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.962893 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/261dc564-2126-415f-a5ce-a988af8a053c-test-operator-ephemeral-workdir\") pod \"261dc564-2126-415f-a5ce-a988af8a053c\" (UID: \"261dc564-2126-415f-a5ce-a988af8a053c\") " Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.962708 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/261dc564-2126-415f-a5ce-a988af8a053c-config-data" (OuterVolumeSpecName: "config-data") pod "261dc564-2126-415f-a5ce-a988af8a053c" (UID: "261dc564-2126-415f-a5ce-a988af8a053c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.969462 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/261dc564-2126-415f-a5ce-a988af8a053c-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "261dc564-2126-415f-a5ce-a988af8a053c" (UID: "261dc564-2126-415f-a5ce-a988af8a053c"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.976642 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/261dc564-2126-415f-a5ce-a988af8a053c-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "261dc564-2126-415f-a5ce-a988af8a053c" (UID: "261dc564-2126-415f-a5ce-a988af8a053c"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.984433 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "261dc564-2126-415f-a5ce-a988af8a053c" (UID: "261dc564-2126-415f-a5ce-a988af8a053c"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:58:46 crc kubenswrapper[4726]: I1001 16:58:46.993786 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/261dc564-2126-415f-a5ce-a988af8a053c-kube-api-access-tmlxv" (OuterVolumeSpecName: "kube-api-access-tmlxv") pod "261dc564-2126-415f-a5ce-a988af8a053c" (UID: "261dc564-2126-415f-a5ce-a988af8a053c"). InnerVolumeSpecName "kube-api-access-tmlxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.014379 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "261dc564-2126-415f-a5ce-a988af8a053c" (UID: "261dc564-2126-415f-a5ce-a988af8a053c"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.032890 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "261dc564-2126-415f-a5ce-a988af8a053c" (UID: "261dc564-2126-415f-a5ce-a988af8a053c"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.043603 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "261dc564-2126-415f-a5ce-a988af8a053c" (UID: "261dc564-2126-415f-a5ce-a988af8a053c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.050423 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/261dc564-2126-415f-a5ce-a988af8a053c-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "261dc564-2126-415f-a5ce-a988af8a053c" (UID: "261dc564-2126-415f-a5ce-a988af8a053c"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.065435 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/261dc564-2126-415f-a5ce-a988af8a053c-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.065504 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.065518 4726 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/261dc564-2126-415f-a5ce-a988af8a053c-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.065548 4726 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/261dc564-2126-415f-a5ce-a988af8a053c-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.065558 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.065567 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/261dc564-2126-415f-a5ce-a988af8a053c-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.065575 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.065582 4726 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/261dc564-2126-415f-a5ce-a988af8a053c-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.065590 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmlxv\" (UniqueName: \"kubernetes.io/projected/261dc564-2126-415f-a5ce-a988af8a053c-kube-api-access-tmlxv\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.084980 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.168091 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.452007 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"261dc564-2126-415f-a5ce-a988af8a053c","Type":"ContainerDied","Data":"265510c1f80135e34f9959c1824ad7a905af7cbcbf00cdef7d2ec67c76500592"} Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.452096 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="265510c1f80135e34f9959c1824ad7a905af7cbcbf00cdef7d2ec67c76500592" Oct 01 16:58:47 crc kubenswrapper[4726]: I1001 16:58:47.452118 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.835549 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 01 16:58:50 crc kubenswrapper[4726]: E1001 16:58:50.836796 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0bac0b4-1db2-4f8f-a226-771135977a8b" containerName="registry-server" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.836818 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0bac0b4-1db2-4f8f-a226-771135977a8b" containerName="registry-server" Oct 01 16:58:50 crc kubenswrapper[4726]: E1001 16:58:50.836844 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="261dc564-2126-415f-a5ce-a988af8a053c" containerName="tempest-tests-tempest-tests-runner" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.836856 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="261dc564-2126-415f-a5ce-a988af8a053c" containerName="tempest-tests-tempest-tests-runner" Oct 01 16:58:50 crc kubenswrapper[4726]: E1001 16:58:50.836873 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54e641e4-4402-4da7-aa1c-159714850518" containerName="extract-content" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.836891 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="54e641e4-4402-4da7-aa1c-159714850518" containerName="extract-content" Oct 01 16:58:50 crc kubenswrapper[4726]: E1001 16:58:50.836904 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0bac0b4-1db2-4f8f-a226-771135977a8b" containerName="extract-content" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.836916 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0bac0b4-1db2-4f8f-a226-771135977a8b" containerName="extract-content" Oct 01 16:58:50 crc kubenswrapper[4726]: E1001 16:58:50.836962 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54e641e4-4402-4da7-aa1c-159714850518" containerName="registry-server" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.836974 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="54e641e4-4402-4da7-aa1c-159714850518" containerName="registry-server" Oct 01 16:58:50 crc kubenswrapper[4726]: E1001 16:58:50.837001 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54e641e4-4402-4da7-aa1c-159714850518" containerName="extract-utilities" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.837014 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="54e641e4-4402-4da7-aa1c-159714850518" containerName="extract-utilities" Oct 01 16:58:50 crc kubenswrapper[4726]: E1001 16:58:50.837074 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0bac0b4-1db2-4f8f-a226-771135977a8b" containerName="extract-utilities" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.837087 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0bac0b4-1db2-4f8f-a226-771135977a8b" containerName="extract-utilities" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.837399 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0bac0b4-1db2-4f8f-a226-771135977a8b" containerName="registry-server" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.837445 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="261dc564-2126-415f-a5ce-a988af8a053c" containerName="tempest-tests-tempest-tests-runner" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.837469 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="54e641e4-4402-4da7-aa1c-159714850518" containerName="registry-server" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.838637 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.841572 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-t9dds" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.850030 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.953212 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c178394b-ef0f-48ec-992f-58cb68ef7ca6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:58:50 crc kubenswrapper[4726]: I1001 16:58:50.953989 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj5nn\" (UniqueName: \"kubernetes.io/projected/c178394b-ef0f-48ec-992f-58cb68ef7ca6-kube-api-access-wj5nn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c178394b-ef0f-48ec-992f-58cb68ef7ca6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:58:51 crc kubenswrapper[4726]: I1001 16:58:51.058162 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c178394b-ef0f-48ec-992f-58cb68ef7ca6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:58:51 crc kubenswrapper[4726]: I1001 16:58:51.058586 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj5nn\" (UniqueName: \"kubernetes.io/projected/c178394b-ef0f-48ec-992f-58cb68ef7ca6-kube-api-access-wj5nn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c178394b-ef0f-48ec-992f-58cb68ef7ca6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:58:51 crc kubenswrapper[4726]: I1001 16:58:51.058868 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c178394b-ef0f-48ec-992f-58cb68ef7ca6\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:58:51 crc kubenswrapper[4726]: I1001 16:58:51.081337 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj5nn\" (UniqueName: \"kubernetes.io/projected/c178394b-ef0f-48ec-992f-58cb68ef7ca6-kube-api-access-wj5nn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c178394b-ef0f-48ec-992f-58cb68ef7ca6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:58:51 crc kubenswrapper[4726]: I1001 16:58:51.096473 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c178394b-ef0f-48ec-992f-58cb68ef7ca6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:58:51 crc kubenswrapper[4726]: I1001 16:58:51.202958 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:58:51 crc kubenswrapper[4726]: I1001 16:58:51.673159 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 01 16:58:52 crc kubenswrapper[4726]: I1001 16:58:52.520180 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"c178394b-ef0f-48ec-992f-58cb68ef7ca6","Type":"ContainerStarted","Data":"dc0d8af4d664763b8d8a2133095439657ce9e4aeaac6a42c43aed79ade685f7f"} Oct 01 16:58:53 crc kubenswrapper[4726]: I1001 16:58:53.413995 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:58:53 crc kubenswrapper[4726]: I1001 16:58:53.414469 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:58:53 crc kubenswrapper[4726]: I1001 16:58:53.536174 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"c178394b-ef0f-48ec-992f-58cb68ef7ca6","Type":"ContainerStarted","Data":"0062d859834f1d6a07fba4aa6142c3c3b17f365f5dde356efb395a8e633e1da2"} Oct 01 16:58:53 crc kubenswrapper[4726]: I1001 16:58:53.566027 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.745110553 podStartE2EDuration="3.565997557s" podCreationTimestamp="2025-10-01 16:58:50 +0000 UTC" firstStartedPulling="2025-10-01 16:58:51.681475939 +0000 UTC m=+3584.583028516" lastFinishedPulling="2025-10-01 16:58:52.502362943 +0000 UTC m=+3585.403915520" observedRunningTime="2025-10-01 16:58:53.555231936 +0000 UTC m=+3586.456784543" watchObservedRunningTime="2025-10-01 16:58:53.565997557 +0000 UTC m=+3586.467550174" Oct 01 16:59:10 crc kubenswrapper[4726]: I1001 16:59:10.128120 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jgg2v/must-gather-c94h2"] Oct 01 16:59:10 crc kubenswrapper[4726]: I1001 16:59:10.131286 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/must-gather-c94h2" Oct 01 16:59:10 crc kubenswrapper[4726]: I1001 16:59:10.132565 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-jgg2v"/"default-dockercfg-6p5jj" Oct 01 16:59:10 crc kubenswrapper[4726]: I1001 16:59:10.133973 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jgg2v"/"kube-root-ca.crt" Oct 01 16:59:10 crc kubenswrapper[4726]: I1001 16:59:10.134229 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jgg2v"/"openshift-service-ca.crt" Oct 01 16:59:10 crc kubenswrapper[4726]: I1001 16:59:10.140384 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jgg2v/must-gather-c94h2"] Oct 01 16:59:10 crc kubenswrapper[4726]: I1001 16:59:10.244629 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwkwf\" (UniqueName: \"kubernetes.io/projected/bdd1d196-704d-4ed3-a919-b4f1ec4ce217-kube-api-access-qwkwf\") pod \"must-gather-c94h2\" (UID: \"bdd1d196-704d-4ed3-a919-b4f1ec4ce217\") " pod="openshift-must-gather-jgg2v/must-gather-c94h2" Oct 01 16:59:10 crc kubenswrapper[4726]: I1001 16:59:10.244808 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bdd1d196-704d-4ed3-a919-b4f1ec4ce217-must-gather-output\") pod \"must-gather-c94h2\" (UID: \"bdd1d196-704d-4ed3-a919-b4f1ec4ce217\") " pod="openshift-must-gather-jgg2v/must-gather-c94h2" Oct 01 16:59:10 crc kubenswrapper[4726]: I1001 16:59:10.346768 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwkwf\" (UniqueName: \"kubernetes.io/projected/bdd1d196-704d-4ed3-a919-b4f1ec4ce217-kube-api-access-qwkwf\") pod \"must-gather-c94h2\" (UID: \"bdd1d196-704d-4ed3-a919-b4f1ec4ce217\") " pod="openshift-must-gather-jgg2v/must-gather-c94h2" Oct 01 16:59:10 crc kubenswrapper[4726]: I1001 16:59:10.346861 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bdd1d196-704d-4ed3-a919-b4f1ec4ce217-must-gather-output\") pod \"must-gather-c94h2\" (UID: \"bdd1d196-704d-4ed3-a919-b4f1ec4ce217\") " pod="openshift-must-gather-jgg2v/must-gather-c94h2" Oct 01 16:59:10 crc kubenswrapper[4726]: I1001 16:59:10.347544 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bdd1d196-704d-4ed3-a919-b4f1ec4ce217-must-gather-output\") pod \"must-gather-c94h2\" (UID: \"bdd1d196-704d-4ed3-a919-b4f1ec4ce217\") " pod="openshift-must-gather-jgg2v/must-gather-c94h2" Oct 01 16:59:10 crc kubenswrapper[4726]: I1001 16:59:10.373269 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwkwf\" (UniqueName: \"kubernetes.io/projected/bdd1d196-704d-4ed3-a919-b4f1ec4ce217-kube-api-access-qwkwf\") pod \"must-gather-c94h2\" (UID: \"bdd1d196-704d-4ed3-a919-b4f1ec4ce217\") " pod="openshift-must-gather-jgg2v/must-gather-c94h2" Oct 01 16:59:10 crc kubenswrapper[4726]: I1001 16:59:10.453951 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/must-gather-c94h2" Oct 01 16:59:10 crc kubenswrapper[4726]: I1001 16:59:10.970137 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jgg2v/must-gather-c94h2"] Oct 01 16:59:10 crc kubenswrapper[4726]: W1001 16:59:10.986300 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd1d196_704d_4ed3_a919_b4f1ec4ce217.slice/crio-51e0bb3ac050185f667e3d712b48b487b7c227ff729b370bd676d3829c53512e WatchSource:0}: Error finding container 51e0bb3ac050185f667e3d712b48b487b7c227ff729b370bd676d3829c53512e: Status 404 returned error can't find the container with id 51e0bb3ac050185f667e3d712b48b487b7c227ff729b370bd676d3829c53512e Oct 01 16:59:11 crc kubenswrapper[4726]: I1001 16:59:11.749892 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgg2v/must-gather-c94h2" event={"ID":"bdd1d196-704d-4ed3-a919-b4f1ec4ce217","Type":"ContainerStarted","Data":"51e0bb3ac050185f667e3d712b48b487b7c227ff729b370bd676d3829c53512e"} Oct 01 16:59:17 crc kubenswrapper[4726]: I1001 16:59:17.801097 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgg2v/must-gather-c94h2" event={"ID":"bdd1d196-704d-4ed3-a919-b4f1ec4ce217","Type":"ContainerStarted","Data":"f6cb3e8504569a12ba7d40e115fab94a701a0cc7bd2b6490362ea6a2b9311b2e"} Oct 01 16:59:17 crc kubenswrapper[4726]: I1001 16:59:17.801774 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgg2v/must-gather-c94h2" event={"ID":"bdd1d196-704d-4ed3-a919-b4f1ec4ce217","Type":"ContainerStarted","Data":"7eb6686dfcaa5d8708af187bad76e73c912908a67c0ab1c5c6c5652e2c78c90c"} Oct 01 16:59:17 crc kubenswrapper[4726]: I1001 16:59:17.823352 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jgg2v/must-gather-c94h2" podStartSLOduration=1.621887633 podStartE2EDuration="7.823322563s" podCreationTimestamp="2025-10-01 16:59:10 +0000 UTC" firstStartedPulling="2025-10-01 16:59:10.992312934 +0000 UTC m=+3603.893865511" lastFinishedPulling="2025-10-01 16:59:17.193747834 +0000 UTC m=+3610.095300441" observedRunningTime="2025-10-01 16:59:17.822087017 +0000 UTC m=+3610.723639604" watchObservedRunningTime="2025-10-01 16:59:17.823322563 +0000 UTC m=+3610.724875170" Oct 01 16:59:19 crc kubenswrapper[4726]: E1001 16:59:19.436859 4726 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.217:46228->38.102.83.217:36913: write tcp 38.102.83.217:46228->38.102.83.217:36913: write: broken pipe Oct 01 16:59:20 crc kubenswrapper[4726]: I1001 16:59:20.830229 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jgg2v/crc-debug-556p4"] Oct 01 16:59:20 crc kubenswrapper[4726]: I1001 16:59:20.835858 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/crc-debug-556p4" Oct 01 16:59:20 crc kubenswrapper[4726]: I1001 16:59:20.981711 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cfd58b86-421a-4e51-9b24-8e8c043de832-host\") pod \"crc-debug-556p4\" (UID: \"cfd58b86-421a-4e51-9b24-8e8c043de832\") " pod="openshift-must-gather-jgg2v/crc-debug-556p4" Oct 01 16:59:20 crc kubenswrapper[4726]: I1001 16:59:20.981789 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psgvm\" (UniqueName: \"kubernetes.io/projected/cfd58b86-421a-4e51-9b24-8e8c043de832-kube-api-access-psgvm\") pod \"crc-debug-556p4\" (UID: \"cfd58b86-421a-4e51-9b24-8e8c043de832\") " pod="openshift-must-gather-jgg2v/crc-debug-556p4" Oct 01 16:59:21 crc kubenswrapper[4726]: I1001 16:59:21.083386 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cfd58b86-421a-4e51-9b24-8e8c043de832-host\") pod \"crc-debug-556p4\" (UID: \"cfd58b86-421a-4e51-9b24-8e8c043de832\") " pod="openshift-must-gather-jgg2v/crc-debug-556p4" Oct 01 16:59:21 crc kubenswrapper[4726]: I1001 16:59:21.083449 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psgvm\" (UniqueName: \"kubernetes.io/projected/cfd58b86-421a-4e51-9b24-8e8c043de832-kube-api-access-psgvm\") pod \"crc-debug-556p4\" (UID: \"cfd58b86-421a-4e51-9b24-8e8c043de832\") " pod="openshift-must-gather-jgg2v/crc-debug-556p4" Oct 01 16:59:21 crc kubenswrapper[4726]: I1001 16:59:21.083881 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cfd58b86-421a-4e51-9b24-8e8c043de832-host\") pod \"crc-debug-556p4\" (UID: \"cfd58b86-421a-4e51-9b24-8e8c043de832\") " pod="openshift-must-gather-jgg2v/crc-debug-556p4" Oct 01 16:59:21 crc kubenswrapper[4726]: I1001 16:59:21.102156 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psgvm\" (UniqueName: \"kubernetes.io/projected/cfd58b86-421a-4e51-9b24-8e8c043de832-kube-api-access-psgvm\") pod \"crc-debug-556p4\" (UID: \"cfd58b86-421a-4e51-9b24-8e8c043de832\") " pod="openshift-must-gather-jgg2v/crc-debug-556p4" Oct 01 16:59:21 crc kubenswrapper[4726]: I1001 16:59:21.164945 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/crc-debug-556p4" Oct 01 16:59:21 crc kubenswrapper[4726]: W1001 16:59:21.221129 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfd58b86_421a_4e51_9b24_8e8c043de832.slice/crio-c12c29bd911e00dcca32015d5eee3894ae97e07e179f7ac09d57249fe2d69952 WatchSource:0}: Error finding container c12c29bd911e00dcca32015d5eee3894ae97e07e179f7ac09d57249fe2d69952: Status 404 returned error can't find the container with id c12c29bd911e00dcca32015d5eee3894ae97e07e179f7ac09d57249fe2d69952 Oct 01 16:59:21 crc kubenswrapper[4726]: I1001 16:59:21.843760 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgg2v/crc-debug-556p4" event={"ID":"cfd58b86-421a-4e51-9b24-8e8c043de832","Type":"ContainerStarted","Data":"c12c29bd911e00dcca32015d5eee3894ae97e07e179f7ac09d57249fe2d69952"} Oct 01 16:59:23 crc kubenswrapper[4726]: I1001 16:59:23.413506 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:59:23 crc kubenswrapper[4726]: I1001 16:59:23.413990 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:59:35 crc kubenswrapper[4726]: I1001 16:59:35.993233 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgg2v/crc-debug-556p4" event={"ID":"cfd58b86-421a-4e51-9b24-8e8c043de832","Type":"ContainerStarted","Data":"14c1a933b7e653d9a8218f2bf6a40982eafba5999ff73f7ddb84b1bce7f41d6d"} Oct 01 16:59:36 crc kubenswrapper[4726]: I1001 16:59:36.008356 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jgg2v/crc-debug-556p4" podStartSLOduration=1.81991023 podStartE2EDuration="16.008338905s" podCreationTimestamp="2025-10-01 16:59:20 +0000 UTC" firstStartedPulling="2025-10-01 16:59:21.223588789 +0000 UTC m=+3614.125141366" lastFinishedPulling="2025-10-01 16:59:35.412017424 +0000 UTC m=+3628.313570041" observedRunningTime="2025-10-01 16:59:36.004150884 +0000 UTC m=+3628.905703461" watchObservedRunningTime="2025-10-01 16:59:36.008338905 +0000 UTC m=+3628.909891492" Oct 01 16:59:53 crc kubenswrapper[4726]: I1001 16:59:53.413234 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:59:53 crc kubenswrapper[4726]: I1001 16:59:53.415314 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:59:53 crc kubenswrapper[4726]: I1001 16:59:53.415520 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 16:59:53 crc kubenswrapper[4726]: I1001 16:59:53.416549 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:59:53 crc kubenswrapper[4726]: I1001 16:59:53.416726 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" gracePeriod=600 Oct 01 16:59:53 crc kubenswrapper[4726]: E1001 16:59:53.549843 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 16:59:54 crc kubenswrapper[4726]: I1001 16:59:54.163220 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" exitCode=0 Oct 01 16:59:54 crc kubenswrapper[4726]: I1001 16:59:54.163272 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865"} Oct 01 16:59:54 crc kubenswrapper[4726]: I1001 16:59:54.163309 4726 scope.go:117] "RemoveContainer" containerID="46f893977f5f67930718136271698b5b624cfc7f800e094b51bbecb8e5163132" Oct 01 16:59:54 crc kubenswrapper[4726]: I1001 16:59:54.164380 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 16:59:54 crc kubenswrapper[4726]: E1001 16:59:54.164825 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.178139 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z"] Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.180327 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.182433 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.182500 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.189668 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z"] Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.267834 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5dvj\" (UniqueName: \"kubernetes.io/projected/bced0b30-270e-409f-90e6-9c8c64667c3f-kube-api-access-r5dvj\") pod \"collect-profiles-29322300-mnf4z\" (UID: \"bced0b30-270e-409f-90e6-9c8c64667c3f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.268323 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bced0b30-270e-409f-90e6-9c8c64667c3f-config-volume\") pod \"collect-profiles-29322300-mnf4z\" (UID: \"bced0b30-270e-409f-90e6-9c8c64667c3f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.268355 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bced0b30-270e-409f-90e6-9c8c64667c3f-secret-volume\") pod \"collect-profiles-29322300-mnf4z\" (UID: \"bced0b30-270e-409f-90e6-9c8c64667c3f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.369923 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5dvj\" (UniqueName: \"kubernetes.io/projected/bced0b30-270e-409f-90e6-9c8c64667c3f-kube-api-access-r5dvj\") pod \"collect-profiles-29322300-mnf4z\" (UID: \"bced0b30-270e-409f-90e6-9c8c64667c3f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.370093 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bced0b30-270e-409f-90e6-9c8c64667c3f-config-volume\") pod \"collect-profiles-29322300-mnf4z\" (UID: \"bced0b30-270e-409f-90e6-9c8c64667c3f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.370111 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bced0b30-270e-409f-90e6-9c8c64667c3f-secret-volume\") pod \"collect-profiles-29322300-mnf4z\" (UID: \"bced0b30-270e-409f-90e6-9c8c64667c3f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.371678 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bced0b30-270e-409f-90e6-9c8c64667c3f-config-volume\") pod \"collect-profiles-29322300-mnf4z\" (UID: \"bced0b30-270e-409f-90e6-9c8c64667c3f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.379141 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bced0b30-270e-409f-90e6-9c8c64667c3f-secret-volume\") pod \"collect-profiles-29322300-mnf4z\" (UID: \"bced0b30-270e-409f-90e6-9c8c64667c3f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.407776 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5dvj\" (UniqueName: \"kubernetes.io/projected/bced0b30-270e-409f-90e6-9c8c64667c3f-kube-api-access-r5dvj\") pod \"collect-profiles-29322300-mnf4z\" (UID: \"bced0b30-270e-409f-90e6-9c8c64667c3f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.508813 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" Oct 01 17:00:00 crc kubenswrapper[4726]: I1001 17:00:00.978597 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z"] Oct 01 17:00:01 crc kubenswrapper[4726]: I1001 17:00:01.230219 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" event={"ID":"bced0b30-270e-409f-90e6-9c8c64667c3f","Type":"ContainerStarted","Data":"d21a92c0e04b451430b95ee8d4ff3c936864e9b8e1fe23b0f1981494fdf2300f"} Oct 01 17:00:01 crc kubenswrapper[4726]: I1001 17:00:01.230560 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" event={"ID":"bced0b30-270e-409f-90e6-9c8c64667c3f","Type":"ContainerStarted","Data":"0f1d837ff9a2468db99ca7c13a43e27c45373576ccc3bc59dd169c224490a11e"} Oct 01 17:00:01 crc kubenswrapper[4726]: I1001 17:00:01.247576 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" podStartSLOduration=1.247556001 podStartE2EDuration="1.247556001s" podCreationTimestamp="2025-10-01 17:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 17:00:01.243509755 +0000 UTC m=+3654.145062332" watchObservedRunningTime="2025-10-01 17:00:01.247556001 +0000 UTC m=+3654.149108578" Oct 01 17:00:01 crc kubenswrapper[4726]: E1001 17:00:01.480618 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbced0b30_270e_409f_90e6_9c8c64667c3f.slice/crio-conmon-d21a92c0e04b451430b95ee8d4ff3c936864e9b8e1fe23b0f1981494fdf2300f.scope\": RecentStats: unable to find data in memory cache]" Oct 01 17:00:02 crc kubenswrapper[4726]: I1001 17:00:02.264267 4726 generic.go:334] "Generic (PLEG): container finished" podID="bced0b30-270e-409f-90e6-9c8c64667c3f" containerID="d21a92c0e04b451430b95ee8d4ff3c936864e9b8e1fe23b0f1981494fdf2300f" exitCode=0 Oct 01 17:00:02 crc kubenswrapper[4726]: I1001 17:00:02.264320 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" event={"ID":"bced0b30-270e-409f-90e6-9c8c64667c3f","Type":"ContainerDied","Data":"d21a92c0e04b451430b95ee8d4ff3c936864e9b8e1fe23b0f1981494fdf2300f"} Oct 01 17:00:03 crc kubenswrapper[4726]: I1001 17:00:03.606283 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" Oct 01 17:00:03 crc kubenswrapper[4726]: I1001 17:00:03.741788 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5dvj\" (UniqueName: \"kubernetes.io/projected/bced0b30-270e-409f-90e6-9c8c64667c3f-kube-api-access-r5dvj\") pod \"bced0b30-270e-409f-90e6-9c8c64667c3f\" (UID: \"bced0b30-270e-409f-90e6-9c8c64667c3f\") " Oct 01 17:00:03 crc kubenswrapper[4726]: I1001 17:00:03.741910 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bced0b30-270e-409f-90e6-9c8c64667c3f-config-volume\") pod \"bced0b30-270e-409f-90e6-9c8c64667c3f\" (UID: \"bced0b30-270e-409f-90e6-9c8c64667c3f\") " Oct 01 17:00:03 crc kubenswrapper[4726]: I1001 17:00:03.741996 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bced0b30-270e-409f-90e6-9c8c64667c3f-secret-volume\") pod \"bced0b30-270e-409f-90e6-9c8c64667c3f\" (UID: \"bced0b30-270e-409f-90e6-9c8c64667c3f\") " Oct 01 17:00:03 crc kubenswrapper[4726]: I1001 17:00:03.742910 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bced0b30-270e-409f-90e6-9c8c64667c3f-config-volume" (OuterVolumeSpecName: "config-volume") pod "bced0b30-270e-409f-90e6-9c8c64667c3f" (UID: "bced0b30-270e-409f-90e6-9c8c64667c3f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 17:00:03 crc kubenswrapper[4726]: I1001 17:00:03.749255 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bced0b30-270e-409f-90e6-9c8c64667c3f-kube-api-access-r5dvj" (OuterVolumeSpecName: "kube-api-access-r5dvj") pod "bced0b30-270e-409f-90e6-9c8c64667c3f" (UID: "bced0b30-270e-409f-90e6-9c8c64667c3f"). InnerVolumeSpecName "kube-api-access-r5dvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 17:00:03 crc kubenswrapper[4726]: I1001 17:00:03.753169 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bced0b30-270e-409f-90e6-9c8c64667c3f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bced0b30-270e-409f-90e6-9c8c64667c3f" (UID: "bced0b30-270e-409f-90e6-9c8c64667c3f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 17:00:03 crc kubenswrapper[4726]: I1001 17:00:03.844399 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bced0b30-270e-409f-90e6-9c8c64667c3f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 17:00:03 crc kubenswrapper[4726]: I1001 17:00:03.844435 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5dvj\" (UniqueName: \"kubernetes.io/projected/bced0b30-270e-409f-90e6-9c8c64667c3f-kube-api-access-r5dvj\") on node \"crc\" DevicePath \"\"" Oct 01 17:00:03 crc kubenswrapper[4726]: I1001 17:00:03.844444 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bced0b30-270e-409f-90e6-9c8c64667c3f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 17:00:04 crc kubenswrapper[4726]: I1001 17:00:04.282467 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" event={"ID":"bced0b30-270e-409f-90e6-9c8c64667c3f","Type":"ContainerDied","Data":"0f1d837ff9a2468db99ca7c13a43e27c45373576ccc3bc59dd169c224490a11e"} Oct 01 17:00:04 crc kubenswrapper[4726]: I1001 17:00:04.282798 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f1d837ff9a2468db99ca7c13a43e27c45373576ccc3bc59dd169c224490a11e" Oct 01 17:00:04 crc kubenswrapper[4726]: I1001 17:00:04.282557 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322300-mnf4z" Oct 01 17:00:04 crc kubenswrapper[4726]: I1001 17:00:04.318926 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q"] Oct 01 17:00:04 crc kubenswrapper[4726]: I1001 17:00:04.327038 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322255-tgl6q"] Oct 01 17:00:05 crc kubenswrapper[4726]: I1001 17:00:05.822423 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b99038f-6158-4768-ad4e-b6e9cec7bf72" path="/var/lib/kubelet/pods/0b99038f-6158-4768-ad4e-b6e9cec7bf72/volumes" Oct 01 17:00:07 crc kubenswrapper[4726]: I1001 17:00:07.816033 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:00:07 crc kubenswrapper[4726]: E1001 17:00:07.816612 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:00:18 crc kubenswrapper[4726]: I1001 17:00:18.808529 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:00:18 crc kubenswrapper[4726]: E1001 17:00:18.809222 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:00:22 crc kubenswrapper[4726]: I1001 17:00:22.206032 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5649c4d89d-cjbv9_e28be6af-1011-49d4-a949-77fdf75e07d9/barbican-api/0.log" Oct 01 17:00:22 crc kubenswrapper[4726]: I1001 17:00:22.268467 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5649c4d89d-cjbv9_e28be6af-1011-49d4-a949-77fdf75e07d9/barbican-api-log/0.log" Oct 01 17:00:22 crc kubenswrapper[4726]: I1001 17:00:22.400953 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-99545b666-4wmpl_7f25ba81-3f98-4d63-8035-2b30a1df5f54/barbican-keystone-listener/0.log" Oct 01 17:00:22 crc kubenswrapper[4726]: I1001 17:00:22.485510 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-99545b666-4wmpl_7f25ba81-3f98-4d63-8035-2b30a1df5f54/barbican-keystone-listener-log/0.log" Oct 01 17:00:22 crc kubenswrapper[4726]: I1001 17:00:22.685999 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-57cbb4948c-gcc56_94005d17-9bb0-4039-b493-5057a554f054/barbican-worker/0.log" Oct 01 17:00:22 crc kubenswrapper[4726]: I1001 17:00:22.718633 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-57cbb4948c-gcc56_94005d17-9bb0-4039-b493-5057a554f054/barbican-worker-log/0.log" Oct 01 17:00:23 crc kubenswrapper[4726]: I1001 17:00:23.033040 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd_bf9f4ecc-09d0-482c-81d2-d7fe9503de1f/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:23 crc kubenswrapper[4726]: I1001 17:00:23.220884 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a303d6a-38ee-4641-b10b-d2627121df03/ceilometer-central-agent/0.log" Oct 01 17:00:23 crc kubenswrapper[4726]: I1001 17:00:23.285025 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a303d6a-38ee-4641-b10b-d2627121df03/proxy-httpd/0.log" Oct 01 17:00:23 crc kubenswrapper[4726]: I1001 17:00:23.287141 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a303d6a-38ee-4641-b10b-d2627121df03/ceilometer-notification-agent/0.log" Oct 01 17:00:23 crc kubenswrapper[4726]: I1001 17:00:23.434969 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a303d6a-38ee-4641-b10b-d2627121df03/sg-core/0.log" Oct 01 17:00:23 crc kubenswrapper[4726]: I1001 17:00:23.490971 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7/cinder-api/0.log" Oct 01 17:00:23 crc kubenswrapper[4726]: I1001 17:00:23.659796 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7/cinder-api-log/0.log" Oct 01 17:00:23 crc kubenswrapper[4726]: I1001 17:00:23.742904 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_aae75086-c497-417c-82aa-9ed0e320b325/cinder-scheduler/0.log" Oct 01 17:00:23 crc kubenswrapper[4726]: I1001 17:00:23.963380 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_aae75086-c497-417c-82aa-9ed0e320b325/probe/0.log" Oct 01 17:00:24 crc kubenswrapper[4726]: I1001 17:00:24.071650 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-fttzz_ec513d51-0278-4d26-aaa7-a66264c32356/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:24 crc kubenswrapper[4726]: I1001 17:00:24.450679 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-4s76w_b8c389e9-bf2c-4e85-a2ec-fc334262347f/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:24 crc kubenswrapper[4726]: I1001 17:00:24.557476 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk_98f0cdbd-7ea0-486b-92bc-14e78ab54ffd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:24 crc kubenswrapper[4726]: I1001 17:00:24.753141 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-cx7ps_09379814-0816-4c10-ba61-5d2875e3841a/init/0.log" Oct 01 17:00:24 crc kubenswrapper[4726]: I1001 17:00:24.927427 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-cx7ps_09379814-0816-4c10-ba61-5d2875e3841a/init/0.log" Oct 01 17:00:24 crc kubenswrapper[4726]: I1001 17:00:24.966765 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-cx7ps_09379814-0816-4c10-ba61-5d2875e3841a/dnsmasq-dns/0.log" Oct 01 17:00:25 crc kubenswrapper[4726]: I1001 17:00:25.127134 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-g7c58_bf27af63-5f62-4d0b-98e6-a7a42fb80d7b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:25 crc kubenswrapper[4726]: I1001 17:00:25.214942 4726 scope.go:117] "RemoveContainer" containerID="17d9a341b2fae3ddffbbbd1b780de16a97302f2fb7fea899d568340653c1bfa1" Oct 01 17:00:25 crc kubenswrapper[4726]: I1001 17:00:25.245007 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_7aa87400-a85d-4d53-8b65-915138d87c01/glance-httpd/0.log" Oct 01 17:00:25 crc kubenswrapper[4726]: I1001 17:00:25.359187 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_7aa87400-a85d-4d53-8b65-915138d87c01/glance-log/0.log" Oct 01 17:00:25 crc kubenswrapper[4726]: I1001 17:00:25.503497 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b4a301ba-9e78-4262-9efa-3d4d91a5957a/glance-httpd/0.log" Oct 01 17:00:25 crc kubenswrapper[4726]: I1001 17:00:25.562697 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b4a301ba-9e78-4262-9efa-3d4d91a5957a/glance-log/0.log" Oct 01 17:00:25 crc kubenswrapper[4726]: I1001 17:00:25.908816 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7b5b89f8d4-d8p9g_c8ee9afe-8056-4de3-bbff-524420e36290/horizon/0.log" Oct 01 17:00:25 crc kubenswrapper[4726]: I1001 17:00:25.975383 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j_d6a1971b-c5b2-4a12-90f2-d986c3ac37b2/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:26 crc kubenswrapper[4726]: I1001 17:00:26.122204 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7b5b89f8d4-d8p9g_c8ee9afe-8056-4de3-bbff-524420e36290/horizon-log/0.log" Oct 01 17:00:26 crc kubenswrapper[4726]: I1001 17:00:26.144665 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-qrzm2_4cdcd93d-244a-4adb-b295-ae609306b749/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:26 crc kubenswrapper[4726]: I1001 17:00:26.364813 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-75bf4bdcc5-mbcpq_033ea7d7-f7e9-49dd-8c50-5402d4a1aa89/keystone-api/0.log" Oct 01 17:00:26 crc kubenswrapper[4726]: I1001 17:00:26.385254 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_d17a343c-2ff2-4c6e-a224-c8d13f476243/kube-state-metrics/0.log" Oct 01 17:00:26 crc kubenswrapper[4726]: I1001 17:00:26.573253 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-svtfg_a66fccbb-d629-4290-a469-066027643cfa/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:26 crc kubenswrapper[4726]: I1001 17:00:26.928653 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-598fc759b9-28cln_16ed9afc-f760-4a59-8db3-8d76c1bd42fa/neutron-api/0.log" Oct 01 17:00:26 crc kubenswrapper[4726]: I1001 17:00:26.980951 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-598fc759b9-28cln_16ed9afc-f760-4a59-8db3-8d76c1bd42fa/neutron-httpd/0.log" Oct 01 17:00:27 crc kubenswrapper[4726]: I1001 17:00:27.183024 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6_ef118af6-8d88-4bdd-bb43-c8532cd50f42/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:27 crc kubenswrapper[4726]: I1001 17:00:27.777134 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6188ef34-0980-4918-bb75-27886226804e/nova-api-log/0.log" Oct 01 17:00:27 crc kubenswrapper[4726]: I1001 17:00:27.892529 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6188ef34-0980-4918-bb75-27886226804e/nova-api-api/0.log" Oct 01 17:00:27 crc kubenswrapper[4726]: I1001 17:00:27.942746 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_db204841-b5e7-4ed3-bd88-e6fb7f9d627f/nova-cell0-conductor-conductor/0.log" Oct 01 17:00:28 crc kubenswrapper[4726]: I1001 17:00:28.336259 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_812fbec4-8c87-4e63-b7e6-e165412dc519/nova-cell1-conductor-conductor/0.log" Oct 01 17:00:28 crc kubenswrapper[4726]: I1001 17:00:28.617830 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_0c2bafde-fa7b-4e17-a5da-e2f2caa63012/nova-cell1-novncproxy-novncproxy/0.log" Oct 01 17:00:28 crc kubenswrapper[4726]: I1001 17:00:28.659358 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-d77vq_d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:28 crc kubenswrapper[4726]: I1001 17:00:28.993310 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8ec3f7f2-1bb9-47a0-9f62-54625de5bf67/nova-metadata-log/0.log" Oct 01 17:00:29 crc kubenswrapper[4726]: I1001 17:00:29.433475 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c094983e-ab3f-4359-a0ed-5481f349409a/nova-scheduler-scheduler/0.log" Oct 01 17:00:29 crc kubenswrapper[4726]: I1001 17:00:29.532024 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e08387ed-ffb4-4488-9345-02504e1c66cc/mysql-bootstrap/0.log" Oct 01 17:00:29 crc kubenswrapper[4726]: I1001 17:00:29.770208 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e08387ed-ffb4-4488-9345-02504e1c66cc/mysql-bootstrap/0.log" Oct 01 17:00:29 crc kubenswrapper[4726]: I1001 17:00:29.778992 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e08387ed-ffb4-4488-9345-02504e1c66cc/galera/0.log" Oct 01 17:00:29 crc kubenswrapper[4726]: I1001 17:00:29.813540 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:00:29 crc kubenswrapper[4726]: E1001 17:00:29.813742 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:00:30 crc kubenswrapper[4726]: I1001 17:00:30.054599 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d743678c-e92d-4515-88d0-27e4a5086aac/mysql-bootstrap/0.log" Oct 01 17:00:30 crc kubenswrapper[4726]: I1001 17:00:30.225002 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d743678c-e92d-4515-88d0-27e4a5086aac/mysql-bootstrap/0.log" Oct 01 17:00:30 crc kubenswrapper[4726]: I1001 17:00:30.253855 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8ec3f7f2-1bb9-47a0-9f62-54625de5bf67/nova-metadata-metadata/0.log" Oct 01 17:00:30 crc kubenswrapper[4726]: I1001 17:00:30.293724 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d743678c-e92d-4515-88d0-27e4a5086aac/galera/0.log" Oct 01 17:00:30 crc kubenswrapper[4726]: I1001 17:00:30.483179 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_37c371f5-1756-49e1-82ec-40b7cb86f524/openstackclient/0.log" Oct 01 17:00:30 crc kubenswrapper[4726]: I1001 17:00:30.650297 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-h598x_72fa255e-484d-43ab-a837-d31c20e8e4d4/openstack-network-exporter/0.log" Oct 01 17:00:30 crc kubenswrapper[4726]: I1001 17:00:30.900312 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4msxz_525c9172-edd7-4483-b649-128a2d7ac283/ovsdb-server-init/0.log" Oct 01 17:00:31 crc kubenswrapper[4726]: I1001 17:00:31.097746 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4msxz_525c9172-edd7-4483-b649-128a2d7ac283/ovsdb-server-init/0.log" Oct 01 17:00:31 crc kubenswrapper[4726]: I1001 17:00:31.129635 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4msxz_525c9172-edd7-4483-b649-128a2d7ac283/ovsdb-server/0.log" Oct 01 17:00:31 crc kubenswrapper[4726]: I1001 17:00:31.133765 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4msxz_525c9172-edd7-4483-b649-128a2d7ac283/ovs-vswitchd/0.log" Oct 01 17:00:31 crc kubenswrapper[4726]: I1001 17:00:31.421793 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-z4lk4_f67ec6c1-adf7-47f9-855f-fde38e0430c7/ovn-controller/0.log" Oct 01 17:00:31 crc kubenswrapper[4726]: I1001 17:00:31.562456 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-ptlg9_4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:31 crc kubenswrapper[4726]: I1001 17:00:31.700997 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_cb9896de-b781-451d-8207-093218e40684/openstack-network-exporter/0.log" Oct 01 17:00:31 crc kubenswrapper[4726]: I1001 17:00:31.792855 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_cb9896de-b781-451d-8207-093218e40684/ovn-northd/0.log" Oct 01 17:00:31 crc kubenswrapper[4726]: I1001 17:00:31.977453 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab/openstack-network-exporter/0.log" Oct 01 17:00:32 crc kubenswrapper[4726]: I1001 17:00:32.026394 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab/ovsdbserver-nb/0.log" Oct 01 17:00:32 crc kubenswrapper[4726]: I1001 17:00:32.212996 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_23600bcf-22d0-4de1-be32-5adf413748b4/openstack-network-exporter/0.log" Oct 01 17:00:32 crc kubenswrapper[4726]: I1001 17:00:32.249337 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_23600bcf-22d0-4de1-be32-5adf413748b4/ovsdbserver-sb/0.log" Oct 01 17:00:32 crc kubenswrapper[4726]: I1001 17:00:32.507291 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-68f479c67b-gj9j5_f3ada4b2-6da4-4139-ad97-0d45034918aa/placement-api/0.log" Oct 01 17:00:32 crc kubenswrapper[4726]: I1001 17:00:32.540123 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-68f479c67b-gj9j5_f3ada4b2-6da4-4139-ad97-0d45034918aa/placement-log/0.log" Oct 01 17:00:32 crc kubenswrapper[4726]: I1001 17:00:32.695280 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8885e9d5-6b72-496e-b395-d0f94b3b991c/setup-container/0.log" Oct 01 17:00:32 crc kubenswrapper[4726]: I1001 17:00:32.984314 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8885e9d5-6b72-496e-b395-d0f94b3b991c/setup-container/0.log" Oct 01 17:00:33 crc kubenswrapper[4726]: I1001 17:00:33.023338 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8885e9d5-6b72-496e-b395-d0f94b3b991c/rabbitmq/0.log" Oct 01 17:00:33 crc kubenswrapper[4726]: I1001 17:00:33.185941 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b5c68163-e0cf-43e8-8659-260fb0aca643/setup-container/0.log" Oct 01 17:00:33 crc kubenswrapper[4726]: I1001 17:00:33.435227 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b5c68163-e0cf-43e8-8659-260fb0aca643/setup-container/0.log" Oct 01 17:00:33 crc kubenswrapper[4726]: I1001 17:00:33.465521 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b5c68163-e0cf-43e8-8659-260fb0aca643/rabbitmq/0.log" Oct 01 17:00:33 crc kubenswrapper[4726]: I1001 17:00:33.637395 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c_016c122e-b1fa-43a1-a806-93533e24b8f6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:33 crc kubenswrapper[4726]: I1001 17:00:33.676720 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-rsl2g_fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:33 crc kubenswrapper[4726]: I1001 17:00:33.851109 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn_01436a04-eda0-42cf-a799-29d2bddd197e/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:34 crc kubenswrapper[4726]: I1001 17:00:34.127501 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-6fb8b_1537a5c5-290d-4e30-ba28-f2b90d0d1fe3/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:34 crc kubenswrapper[4726]: I1001 17:00:34.364624 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-mpm5k_e03d54c6-e53e-4837-9d0d-a9183775699e/ssh-known-hosts-edpm-deployment/0.log" Oct 01 17:00:34 crc kubenswrapper[4726]: I1001 17:00:34.503706 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-57c9dff847-vbzrr_85ac310a-31b1-40c9-9724-2c488c0061e2/proxy-httpd/0.log" Oct 01 17:00:34 crc kubenswrapper[4726]: I1001 17:00:34.511606 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-57c9dff847-vbzrr_85ac310a-31b1-40c9-9724-2c488c0061e2/proxy-server/0.log" Oct 01 17:00:34 crc kubenswrapper[4726]: I1001 17:00:34.709840 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-t4hqx_4ab0d371-af59-407e-b596-3c533ad031b8/swift-ring-rebalance/0.log" Oct 01 17:00:34 crc kubenswrapper[4726]: I1001 17:00:34.830310 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/account-auditor/0.log" Oct 01 17:00:34 crc kubenswrapper[4726]: I1001 17:00:34.939132 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/account-reaper/0.log" Oct 01 17:00:34 crc kubenswrapper[4726]: I1001 17:00:34.970912 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/account-replicator/0.log" Oct 01 17:00:35 crc kubenswrapper[4726]: I1001 17:00:35.053441 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/account-server/0.log" Oct 01 17:00:35 crc kubenswrapper[4726]: I1001 17:00:35.127144 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/container-auditor/0.log" Oct 01 17:00:35 crc kubenswrapper[4726]: I1001 17:00:35.218600 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/container-replicator/0.log" Oct 01 17:00:35 crc kubenswrapper[4726]: I1001 17:00:35.273203 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/container-server/0.log" Oct 01 17:00:35 crc kubenswrapper[4726]: I1001 17:00:35.354738 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/container-updater/0.log" Oct 01 17:00:35 crc kubenswrapper[4726]: I1001 17:00:35.443019 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/object-auditor/0.log" Oct 01 17:00:35 crc kubenswrapper[4726]: I1001 17:00:35.495142 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/object-expirer/0.log" Oct 01 17:00:35 crc kubenswrapper[4726]: I1001 17:00:35.625620 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/object-replicator/0.log" Oct 01 17:00:35 crc kubenswrapper[4726]: I1001 17:00:35.662564 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/object-server/0.log" Oct 01 17:00:35 crc kubenswrapper[4726]: I1001 17:00:35.833677 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/object-updater/0.log" Oct 01 17:00:35 crc kubenswrapper[4726]: I1001 17:00:35.837119 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/rsync/0.log" Oct 01 17:00:35 crc kubenswrapper[4726]: I1001 17:00:35.892407 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/swift-recon-cron/0.log" Oct 01 17:00:36 crc kubenswrapper[4726]: I1001 17:00:36.163531 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k_376afdd4-d312-464e-9627-42656734a785/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:36 crc kubenswrapper[4726]: I1001 17:00:36.298747 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_261dc564-2126-415f-a5ce-a988af8a053c/tempest-tests-tempest-tests-runner/0.log" Oct 01 17:00:36 crc kubenswrapper[4726]: I1001 17:00:36.387629 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_c178394b-ef0f-48ec-992f-58cb68ef7ca6/test-operator-logs-container/0.log" Oct 01 17:00:36 crc kubenswrapper[4726]: I1001 17:00:36.587897 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk_c69a5dd3-367d-4db8-907c-cb7afb6da6a9/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:00:43 crc kubenswrapper[4726]: I1001 17:00:43.812359 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:00:43 crc kubenswrapper[4726]: E1001 17:00:43.813042 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:00:44 crc kubenswrapper[4726]: I1001 17:00:44.272287 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fd24681e-73eb-49d0-bd62-4d11a485c18e/memcached/0.log" Oct 01 17:00:57 crc kubenswrapper[4726]: I1001 17:00:57.813912 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:00:57 crc kubenswrapper[4726]: E1001 17:00:57.814548 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.213499 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29322301-t47c4"] Oct 01 17:01:00 crc kubenswrapper[4726]: E1001 17:01:00.215168 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bced0b30-270e-409f-90e6-9c8c64667c3f" containerName="collect-profiles" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.215288 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="bced0b30-270e-409f-90e6-9c8c64667c3f" containerName="collect-profiles" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.215568 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="bced0b30-270e-409f-90e6-9c8c64667c3f" containerName="collect-profiles" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.216379 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.228028 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29322301-t47c4"] Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.293814 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr2pj\" (UniqueName: \"kubernetes.io/projected/91502936-cb39-48a4-9e82-8cec0e8fb5ab-kube-api-access-nr2pj\") pod \"keystone-cron-29322301-t47c4\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.293875 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-fernet-keys\") pod \"keystone-cron-29322301-t47c4\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.293899 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-config-data\") pod \"keystone-cron-29322301-t47c4\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.293923 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-combined-ca-bundle\") pod \"keystone-cron-29322301-t47c4\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.395406 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr2pj\" (UniqueName: \"kubernetes.io/projected/91502936-cb39-48a4-9e82-8cec0e8fb5ab-kube-api-access-nr2pj\") pod \"keystone-cron-29322301-t47c4\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.395477 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-config-data\") pod \"keystone-cron-29322301-t47c4\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.395502 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-fernet-keys\") pod \"keystone-cron-29322301-t47c4\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.395527 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-combined-ca-bundle\") pod \"keystone-cron-29322301-t47c4\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.407245 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-fernet-keys\") pod \"keystone-cron-29322301-t47c4\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.408918 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-config-data\") pod \"keystone-cron-29322301-t47c4\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.412321 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-combined-ca-bundle\") pod \"keystone-cron-29322301-t47c4\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.413718 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr2pj\" (UniqueName: \"kubernetes.io/projected/91502936-cb39-48a4-9e82-8cec0e8fb5ab-kube-api-access-nr2pj\") pod \"keystone-cron-29322301-t47c4\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:00 crc kubenswrapper[4726]: I1001 17:01:00.537263 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:01 crc kubenswrapper[4726]: I1001 17:01:01.016656 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29322301-t47c4"] Oct 01 17:01:01 crc kubenswrapper[4726]: I1001 17:01:01.878669 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322301-t47c4" event={"ID":"91502936-cb39-48a4-9e82-8cec0e8fb5ab","Type":"ContainerStarted","Data":"bf4834161997c6b96f521d1f475e32437ea3e05850a0c3ca262b14f62ace1261"} Oct 01 17:01:01 crc kubenswrapper[4726]: I1001 17:01:01.879057 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322301-t47c4" event={"ID":"91502936-cb39-48a4-9e82-8cec0e8fb5ab","Type":"ContainerStarted","Data":"33badd85162771b4ed0aa05ce7fd38bc4de14a2a8a93a6d6579ce942af9b03cd"} Oct 01 17:01:01 crc kubenswrapper[4726]: I1001 17:01:01.908575 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29322301-t47c4" podStartSLOduration=1.908551023 podStartE2EDuration="1.908551023s" podCreationTimestamp="2025-10-01 17:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 17:01:01.900901273 +0000 UTC m=+3714.802453850" watchObservedRunningTime="2025-10-01 17:01:01.908551023 +0000 UTC m=+3714.810103600" Oct 01 17:01:03 crc kubenswrapper[4726]: I1001 17:01:03.899569 4726 generic.go:334] "Generic (PLEG): container finished" podID="91502936-cb39-48a4-9e82-8cec0e8fb5ab" containerID="bf4834161997c6b96f521d1f475e32437ea3e05850a0c3ca262b14f62ace1261" exitCode=0 Oct 01 17:01:03 crc kubenswrapper[4726]: I1001 17:01:03.899599 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322301-t47c4" event={"ID":"91502936-cb39-48a4-9e82-8cec0e8fb5ab","Type":"ContainerDied","Data":"bf4834161997c6b96f521d1f475e32437ea3e05850a0c3ca262b14f62ace1261"} Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.322081 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.401559 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-combined-ca-bundle\") pod \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.401768 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-fernet-keys\") pod \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.401904 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nr2pj\" (UniqueName: \"kubernetes.io/projected/91502936-cb39-48a4-9e82-8cec0e8fb5ab-kube-api-access-nr2pj\") pod \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.402004 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-config-data\") pod \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\" (UID: \"91502936-cb39-48a4-9e82-8cec0e8fb5ab\") " Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.412441 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "91502936-cb39-48a4-9e82-8cec0e8fb5ab" (UID: "91502936-cb39-48a4-9e82-8cec0e8fb5ab"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.412486 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91502936-cb39-48a4-9e82-8cec0e8fb5ab-kube-api-access-nr2pj" (OuterVolumeSpecName: "kube-api-access-nr2pj") pod "91502936-cb39-48a4-9e82-8cec0e8fb5ab" (UID: "91502936-cb39-48a4-9e82-8cec0e8fb5ab"). InnerVolumeSpecName "kube-api-access-nr2pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.438668 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91502936-cb39-48a4-9e82-8cec0e8fb5ab" (UID: "91502936-cb39-48a4-9e82-8cec0e8fb5ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.487338 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-config-data" (OuterVolumeSpecName: "config-data") pod "91502936-cb39-48a4-9e82-8cec0e8fb5ab" (UID: "91502936-cb39-48a4-9e82-8cec0e8fb5ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.506395 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.506460 4726 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.506489 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nr2pj\" (UniqueName: \"kubernetes.io/projected/91502936-cb39-48a4-9e82-8cec0e8fb5ab-kube-api-access-nr2pj\") on node \"crc\" DevicePath \"\"" Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.506514 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91502936-cb39-48a4-9e82-8cec0e8fb5ab-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.918426 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322301-t47c4" event={"ID":"91502936-cb39-48a4-9e82-8cec0e8fb5ab","Type":"ContainerDied","Data":"33badd85162771b4ed0aa05ce7fd38bc4de14a2a8a93a6d6579ce942af9b03cd"} Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.918885 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33badd85162771b4ed0aa05ce7fd38bc4de14a2a8a93a6d6579ce942af9b03cd" Oct 01 17:01:05 crc kubenswrapper[4726]: I1001 17:01:05.918634 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322301-t47c4" Oct 01 17:01:08 crc kubenswrapper[4726]: I1001 17:01:08.809163 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:01:08 crc kubenswrapper[4726]: E1001 17:01:08.809863 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:01:22 crc kubenswrapper[4726]: I1001 17:01:22.809727 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:01:22 crc kubenswrapper[4726]: E1001 17:01:22.810925 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:01:33 crc kubenswrapper[4726]: I1001 17:01:33.808099 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:01:33 crc kubenswrapper[4726]: E1001 17:01:33.808673 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:01:41 crc kubenswrapper[4726]: I1001 17:01:41.309948 4726 generic.go:334] "Generic (PLEG): container finished" podID="cfd58b86-421a-4e51-9b24-8e8c043de832" containerID="14c1a933b7e653d9a8218f2bf6a40982eafba5999ff73f7ddb84b1bce7f41d6d" exitCode=0 Oct 01 17:01:41 crc kubenswrapper[4726]: I1001 17:01:41.310037 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgg2v/crc-debug-556p4" event={"ID":"cfd58b86-421a-4e51-9b24-8e8c043de832","Type":"ContainerDied","Data":"14c1a933b7e653d9a8218f2bf6a40982eafba5999ff73f7ddb84b1bce7f41d6d"} Oct 01 17:01:42 crc kubenswrapper[4726]: I1001 17:01:42.472126 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/crc-debug-556p4" Oct 01 17:01:42 crc kubenswrapper[4726]: I1001 17:01:42.522969 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jgg2v/crc-debug-556p4"] Oct 01 17:01:42 crc kubenswrapper[4726]: I1001 17:01:42.536467 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jgg2v/crc-debug-556p4"] Oct 01 17:01:42 crc kubenswrapper[4726]: I1001 17:01:42.605210 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psgvm\" (UniqueName: \"kubernetes.io/projected/cfd58b86-421a-4e51-9b24-8e8c043de832-kube-api-access-psgvm\") pod \"cfd58b86-421a-4e51-9b24-8e8c043de832\" (UID: \"cfd58b86-421a-4e51-9b24-8e8c043de832\") " Oct 01 17:01:42 crc kubenswrapper[4726]: I1001 17:01:42.605433 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cfd58b86-421a-4e51-9b24-8e8c043de832-host\") pod \"cfd58b86-421a-4e51-9b24-8e8c043de832\" (UID: \"cfd58b86-421a-4e51-9b24-8e8c043de832\") " Oct 01 17:01:42 crc kubenswrapper[4726]: I1001 17:01:42.605506 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfd58b86-421a-4e51-9b24-8e8c043de832-host" (OuterVolumeSpecName: "host") pod "cfd58b86-421a-4e51-9b24-8e8c043de832" (UID: "cfd58b86-421a-4e51-9b24-8e8c043de832"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 17:01:42 crc kubenswrapper[4726]: I1001 17:01:42.605848 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cfd58b86-421a-4e51-9b24-8e8c043de832-host\") on node \"crc\" DevicePath \"\"" Oct 01 17:01:42 crc kubenswrapper[4726]: I1001 17:01:42.616320 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfd58b86-421a-4e51-9b24-8e8c043de832-kube-api-access-psgvm" (OuterVolumeSpecName: "kube-api-access-psgvm") pod "cfd58b86-421a-4e51-9b24-8e8c043de832" (UID: "cfd58b86-421a-4e51-9b24-8e8c043de832"). InnerVolumeSpecName "kube-api-access-psgvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 17:01:42 crc kubenswrapper[4726]: I1001 17:01:42.707230 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psgvm\" (UniqueName: \"kubernetes.io/projected/cfd58b86-421a-4e51-9b24-8e8c043de832-kube-api-access-psgvm\") on node \"crc\" DevicePath \"\"" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.337387 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c12c29bd911e00dcca32015d5eee3894ae97e07e179f7ac09d57249fe2d69952" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.337511 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/crc-debug-556p4" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.758010 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jgg2v/crc-debug-45bq5"] Oct 01 17:01:43 crc kubenswrapper[4726]: E1001 17:01:43.758686 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfd58b86-421a-4e51-9b24-8e8c043de832" containerName="container-00" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.758700 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfd58b86-421a-4e51-9b24-8e8c043de832" containerName="container-00" Oct 01 17:01:43 crc kubenswrapper[4726]: E1001 17:01:43.758736 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91502936-cb39-48a4-9e82-8cec0e8fb5ab" containerName="keystone-cron" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.758742 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="91502936-cb39-48a4-9e82-8cec0e8fb5ab" containerName="keystone-cron" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.758947 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="91502936-cb39-48a4-9e82-8cec0e8fb5ab" containerName="keystone-cron" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.758975 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfd58b86-421a-4e51-9b24-8e8c043de832" containerName="container-00" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.759871 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/crc-debug-45bq5" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.823589 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfd58b86-421a-4e51-9b24-8e8c043de832" path="/var/lib/kubelet/pods/cfd58b86-421a-4e51-9b24-8e8c043de832/volumes" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.828003 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4prb8\" (UniqueName: \"kubernetes.io/projected/882435e3-7da3-4144-8fa8-d83002fa8459-kube-api-access-4prb8\") pod \"crc-debug-45bq5\" (UID: \"882435e3-7da3-4144-8fa8-d83002fa8459\") " pod="openshift-must-gather-jgg2v/crc-debug-45bq5" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.828105 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/882435e3-7da3-4144-8fa8-d83002fa8459-host\") pod \"crc-debug-45bq5\" (UID: \"882435e3-7da3-4144-8fa8-d83002fa8459\") " pod="openshift-must-gather-jgg2v/crc-debug-45bq5" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.930010 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4prb8\" (UniqueName: \"kubernetes.io/projected/882435e3-7da3-4144-8fa8-d83002fa8459-kube-api-access-4prb8\") pod \"crc-debug-45bq5\" (UID: \"882435e3-7da3-4144-8fa8-d83002fa8459\") " pod="openshift-must-gather-jgg2v/crc-debug-45bq5" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.930354 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/882435e3-7da3-4144-8fa8-d83002fa8459-host\") pod \"crc-debug-45bq5\" (UID: \"882435e3-7da3-4144-8fa8-d83002fa8459\") " pod="openshift-must-gather-jgg2v/crc-debug-45bq5" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.930667 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/882435e3-7da3-4144-8fa8-d83002fa8459-host\") pod \"crc-debug-45bq5\" (UID: \"882435e3-7da3-4144-8fa8-d83002fa8459\") " pod="openshift-must-gather-jgg2v/crc-debug-45bq5" Oct 01 17:01:43 crc kubenswrapper[4726]: I1001 17:01:43.948550 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4prb8\" (UniqueName: \"kubernetes.io/projected/882435e3-7da3-4144-8fa8-d83002fa8459-kube-api-access-4prb8\") pod \"crc-debug-45bq5\" (UID: \"882435e3-7da3-4144-8fa8-d83002fa8459\") " pod="openshift-must-gather-jgg2v/crc-debug-45bq5" Oct 01 17:01:44 crc kubenswrapper[4726]: I1001 17:01:44.076842 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/crc-debug-45bq5" Oct 01 17:01:44 crc kubenswrapper[4726]: I1001 17:01:44.348429 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgg2v/crc-debug-45bq5" event={"ID":"882435e3-7da3-4144-8fa8-d83002fa8459","Type":"ContainerStarted","Data":"75fee6902f38fda61eaad4e237c590d3ca194ff368270be1e25785a78e4b9779"} Oct 01 17:01:45 crc kubenswrapper[4726]: I1001 17:01:45.359638 4726 generic.go:334] "Generic (PLEG): container finished" podID="882435e3-7da3-4144-8fa8-d83002fa8459" containerID="5aecefd230d7bcfd7df54874dd7991638bf56d76689302a7882ca1c83ef17e13" exitCode=0 Oct 01 17:01:45 crc kubenswrapper[4726]: I1001 17:01:45.359740 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgg2v/crc-debug-45bq5" event={"ID":"882435e3-7da3-4144-8fa8-d83002fa8459","Type":"ContainerDied","Data":"5aecefd230d7bcfd7df54874dd7991638bf56d76689302a7882ca1c83ef17e13"} Oct 01 17:01:46 crc kubenswrapper[4726]: I1001 17:01:46.503003 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/crc-debug-45bq5" Oct 01 17:01:46 crc kubenswrapper[4726]: I1001 17:01:46.596267 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/882435e3-7da3-4144-8fa8-d83002fa8459-host\") pod \"882435e3-7da3-4144-8fa8-d83002fa8459\" (UID: \"882435e3-7da3-4144-8fa8-d83002fa8459\") " Oct 01 17:01:46 crc kubenswrapper[4726]: I1001 17:01:46.596322 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4prb8\" (UniqueName: \"kubernetes.io/projected/882435e3-7da3-4144-8fa8-d83002fa8459-kube-api-access-4prb8\") pod \"882435e3-7da3-4144-8fa8-d83002fa8459\" (UID: \"882435e3-7da3-4144-8fa8-d83002fa8459\") " Oct 01 17:01:46 crc kubenswrapper[4726]: I1001 17:01:46.599132 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/882435e3-7da3-4144-8fa8-d83002fa8459-host" (OuterVolumeSpecName: "host") pod "882435e3-7da3-4144-8fa8-d83002fa8459" (UID: "882435e3-7da3-4144-8fa8-d83002fa8459"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 17:01:46 crc kubenswrapper[4726]: I1001 17:01:46.604547 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/882435e3-7da3-4144-8fa8-d83002fa8459-kube-api-access-4prb8" (OuterVolumeSpecName: "kube-api-access-4prb8") pod "882435e3-7da3-4144-8fa8-d83002fa8459" (UID: "882435e3-7da3-4144-8fa8-d83002fa8459"). InnerVolumeSpecName "kube-api-access-4prb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 17:01:46 crc kubenswrapper[4726]: I1001 17:01:46.698240 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/882435e3-7da3-4144-8fa8-d83002fa8459-host\") on node \"crc\" DevicePath \"\"" Oct 01 17:01:46 crc kubenswrapper[4726]: I1001 17:01:46.698275 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4prb8\" (UniqueName: \"kubernetes.io/projected/882435e3-7da3-4144-8fa8-d83002fa8459-kube-api-access-4prb8\") on node \"crc\" DevicePath \"\"" Oct 01 17:01:47 crc kubenswrapper[4726]: I1001 17:01:47.391273 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgg2v/crc-debug-45bq5" event={"ID":"882435e3-7da3-4144-8fa8-d83002fa8459","Type":"ContainerDied","Data":"75fee6902f38fda61eaad4e237c590d3ca194ff368270be1e25785a78e4b9779"} Oct 01 17:01:47 crc kubenswrapper[4726]: I1001 17:01:47.391537 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75fee6902f38fda61eaad4e237c590d3ca194ff368270be1e25785a78e4b9779" Oct 01 17:01:47 crc kubenswrapper[4726]: I1001 17:01:47.391592 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/crc-debug-45bq5" Oct 01 17:01:47 crc kubenswrapper[4726]: I1001 17:01:47.818157 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:01:47 crc kubenswrapper[4726]: E1001 17:01:47.820889 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:01:51 crc kubenswrapper[4726]: I1001 17:01:51.404963 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jgg2v/crc-debug-45bq5"] Oct 01 17:01:51 crc kubenswrapper[4726]: I1001 17:01:51.413697 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jgg2v/crc-debug-45bq5"] Oct 01 17:01:51 crc kubenswrapper[4726]: I1001 17:01:51.818995 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="882435e3-7da3-4144-8fa8-d83002fa8459" path="/var/lib/kubelet/pods/882435e3-7da3-4144-8fa8-d83002fa8459/volumes" Oct 01 17:01:52 crc kubenswrapper[4726]: I1001 17:01:52.609605 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jgg2v/crc-debug-q4jxc"] Oct 01 17:01:52 crc kubenswrapper[4726]: E1001 17:01:52.611381 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="882435e3-7da3-4144-8fa8-d83002fa8459" containerName="container-00" Oct 01 17:01:52 crc kubenswrapper[4726]: I1001 17:01:52.611468 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="882435e3-7da3-4144-8fa8-d83002fa8459" containerName="container-00" Oct 01 17:01:52 crc kubenswrapper[4726]: I1001 17:01:52.613083 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="882435e3-7da3-4144-8fa8-d83002fa8459" containerName="container-00" Oct 01 17:01:52 crc kubenswrapper[4726]: I1001 17:01:52.615234 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/crc-debug-q4jxc" Oct 01 17:01:52 crc kubenswrapper[4726]: I1001 17:01:52.702836 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/170e63ef-d054-4975-a6c9-6ec4b114ca58-host\") pod \"crc-debug-q4jxc\" (UID: \"170e63ef-d054-4975-a6c9-6ec4b114ca58\") " pod="openshift-must-gather-jgg2v/crc-debug-q4jxc" Oct 01 17:01:52 crc kubenswrapper[4726]: I1001 17:01:52.703292 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljh77\" (UniqueName: \"kubernetes.io/projected/170e63ef-d054-4975-a6c9-6ec4b114ca58-kube-api-access-ljh77\") pod \"crc-debug-q4jxc\" (UID: \"170e63ef-d054-4975-a6c9-6ec4b114ca58\") " pod="openshift-must-gather-jgg2v/crc-debug-q4jxc" Oct 01 17:01:52 crc kubenswrapper[4726]: I1001 17:01:52.804944 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljh77\" (UniqueName: \"kubernetes.io/projected/170e63ef-d054-4975-a6c9-6ec4b114ca58-kube-api-access-ljh77\") pod \"crc-debug-q4jxc\" (UID: \"170e63ef-d054-4975-a6c9-6ec4b114ca58\") " pod="openshift-must-gather-jgg2v/crc-debug-q4jxc" Oct 01 17:01:52 crc kubenswrapper[4726]: I1001 17:01:52.805198 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/170e63ef-d054-4975-a6c9-6ec4b114ca58-host\") pod \"crc-debug-q4jxc\" (UID: \"170e63ef-d054-4975-a6c9-6ec4b114ca58\") " pod="openshift-must-gather-jgg2v/crc-debug-q4jxc" Oct 01 17:01:52 crc kubenswrapper[4726]: I1001 17:01:52.805440 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/170e63ef-d054-4975-a6c9-6ec4b114ca58-host\") pod \"crc-debug-q4jxc\" (UID: \"170e63ef-d054-4975-a6c9-6ec4b114ca58\") " pod="openshift-must-gather-jgg2v/crc-debug-q4jxc" Oct 01 17:01:52 crc kubenswrapper[4726]: I1001 17:01:52.847298 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljh77\" (UniqueName: \"kubernetes.io/projected/170e63ef-d054-4975-a6c9-6ec4b114ca58-kube-api-access-ljh77\") pod \"crc-debug-q4jxc\" (UID: \"170e63ef-d054-4975-a6c9-6ec4b114ca58\") " pod="openshift-must-gather-jgg2v/crc-debug-q4jxc" Oct 01 17:01:52 crc kubenswrapper[4726]: I1001 17:01:52.941944 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/crc-debug-q4jxc" Oct 01 17:01:53 crc kubenswrapper[4726]: I1001 17:01:53.443222 4726 generic.go:334] "Generic (PLEG): container finished" podID="170e63ef-d054-4975-a6c9-6ec4b114ca58" containerID="19b96154d5f00aba7bb02804e15d2696fa8b71498be98f730fa3808d871027ba" exitCode=0 Oct 01 17:01:53 crc kubenswrapper[4726]: I1001 17:01:53.443272 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgg2v/crc-debug-q4jxc" event={"ID":"170e63ef-d054-4975-a6c9-6ec4b114ca58","Type":"ContainerDied","Data":"19b96154d5f00aba7bb02804e15d2696fa8b71498be98f730fa3808d871027ba"} Oct 01 17:01:53 crc kubenswrapper[4726]: I1001 17:01:53.443546 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgg2v/crc-debug-q4jxc" event={"ID":"170e63ef-d054-4975-a6c9-6ec4b114ca58","Type":"ContainerStarted","Data":"70f89ffd6dfd59a9395d520f09e2620e85157ef5ce91e2f65f269a55862f6629"} Oct 01 17:01:53 crc kubenswrapper[4726]: I1001 17:01:53.499573 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jgg2v/crc-debug-q4jxc"] Oct 01 17:01:53 crc kubenswrapper[4726]: I1001 17:01:53.509842 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jgg2v/crc-debug-q4jxc"] Oct 01 17:01:54 crc kubenswrapper[4726]: I1001 17:01:54.569302 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/crc-debug-q4jxc" Oct 01 17:01:54 crc kubenswrapper[4726]: I1001 17:01:54.651086 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljh77\" (UniqueName: \"kubernetes.io/projected/170e63ef-d054-4975-a6c9-6ec4b114ca58-kube-api-access-ljh77\") pod \"170e63ef-d054-4975-a6c9-6ec4b114ca58\" (UID: \"170e63ef-d054-4975-a6c9-6ec4b114ca58\") " Oct 01 17:01:54 crc kubenswrapper[4726]: I1001 17:01:54.651229 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/170e63ef-d054-4975-a6c9-6ec4b114ca58-host\") pod \"170e63ef-d054-4975-a6c9-6ec4b114ca58\" (UID: \"170e63ef-d054-4975-a6c9-6ec4b114ca58\") " Oct 01 17:01:54 crc kubenswrapper[4726]: I1001 17:01:54.651516 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/170e63ef-d054-4975-a6c9-6ec4b114ca58-host" (OuterVolumeSpecName: "host") pod "170e63ef-d054-4975-a6c9-6ec4b114ca58" (UID: "170e63ef-d054-4975-a6c9-6ec4b114ca58"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 17:01:54 crc kubenswrapper[4726]: I1001 17:01:54.652064 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/170e63ef-d054-4975-a6c9-6ec4b114ca58-host\") on node \"crc\" DevicePath \"\"" Oct 01 17:01:54 crc kubenswrapper[4726]: I1001 17:01:54.656847 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/170e63ef-d054-4975-a6c9-6ec4b114ca58-kube-api-access-ljh77" (OuterVolumeSpecName: "kube-api-access-ljh77") pod "170e63ef-d054-4975-a6c9-6ec4b114ca58" (UID: "170e63ef-d054-4975-a6c9-6ec4b114ca58"). InnerVolumeSpecName "kube-api-access-ljh77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 17:01:54 crc kubenswrapper[4726]: I1001 17:01:54.754400 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljh77\" (UniqueName: \"kubernetes.io/projected/170e63ef-d054-4975-a6c9-6ec4b114ca58-kube-api-access-ljh77\") on node \"crc\" DevicePath \"\"" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.060319 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd_6eef09a6-cce6-45d8-ac7e-6b09f74d730a/util/0.log" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.258410 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd_6eef09a6-cce6-45d8-ac7e-6b09f74d730a/pull/0.log" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.258433 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd_6eef09a6-cce6-45d8-ac7e-6b09f74d730a/pull/0.log" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.262793 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd_6eef09a6-cce6-45d8-ac7e-6b09f74d730a/util/0.log" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.467677 4726 scope.go:117] "RemoveContainer" containerID="19b96154d5f00aba7bb02804e15d2696fa8b71498be98f730fa3808d871027ba" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.467745 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/crc-debug-q4jxc" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.469285 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd_6eef09a6-cce6-45d8-ac7e-6b09f74d730a/extract/0.log" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.502529 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd_6eef09a6-cce6-45d8-ac7e-6b09f74d730a/pull/0.log" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.511762 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd_6eef09a6-cce6-45d8-ac7e-6b09f74d730a/util/0.log" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.666468 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-wckhv_85f77f5d-6b74-40b5-b238-d3515b4a685c/kube-rbac-proxy/0.log" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.733687 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-wckhv_85f77f5d-6b74-40b5-b238-d3515b4a685c/manager/0.log" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.742314 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-zxd4h_117d55f2-eee8-454b-9091-3ffffd61c547/kube-rbac-proxy/0.log" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.820517 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="170e63ef-d054-4975-a6c9-6ec4b114ca58" path="/var/lib/kubelet/pods/170e63ef-d054-4975-a6c9-6ec4b114ca58/volumes" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.930237 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-wgkz2_71d9bfcf-89ac-440f-b2b1-1c2969859b0d/manager/0.log" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.934269 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-zxd4h_117d55f2-eee8-454b-9091-3ffffd61c547/manager/0.log" Oct 01 17:01:55 crc kubenswrapper[4726]: I1001 17:01:55.967730 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-wgkz2_71d9bfcf-89ac-440f-b2b1-1c2969859b0d/kube-rbac-proxy/0.log" Oct 01 17:01:56 crc kubenswrapper[4726]: I1001 17:01:56.105739 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-67rj9_daf7d005-ec55-4cdb-91d6-eac465e42e17/kube-rbac-proxy/0.log" Oct 01 17:01:56 crc kubenswrapper[4726]: I1001 17:01:56.222245 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-67rj9_daf7d005-ec55-4cdb-91d6-eac465e42e17/manager/0.log" Oct 01 17:01:56 crc kubenswrapper[4726]: I1001 17:01:56.313760 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-qlb92_f82a5172-6387-4a24-a7e4-a5bd4c90d310/kube-rbac-proxy/0.log" Oct 01 17:01:56 crc kubenswrapper[4726]: I1001 17:01:56.349373 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-qlb92_f82a5172-6387-4a24-a7e4-a5bd4c90d310/manager/0.log" Oct 01 17:01:56 crc kubenswrapper[4726]: I1001 17:01:56.421323 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-zjwvl_39860f7b-9a7f-44b1-8bf4-7a2204092c18/kube-rbac-proxy/0.log" Oct 01 17:01:56 crc kubenswrapper[4726]: I1001 17:01:56.525706 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-zjwvl_39860f7b-9a7f-44b1-8bf4-7a2204092c18/manager/0.log" Oct 01 17:01:56 crc kubenswrapper[4726]: I1001 17:01:56.586685 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-z5lfd_6b63c738-b108-4193-9c9d-11f4eb8227aa/kube-rbac-proxy/0.log" Oct 01 17:01:56 crc kubenswrapper[4726]: I1001 17:01:56.732908 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-z5lfd_6b63c738-b108-4193-9c9d-11f4eb8227aa/manager/0.log" Oct 01 17:01:56 crc kubenswrapper[4726]: I1001 17:01:56.747385 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-cpbbq_8e085c54-533c-4186-a6fe-5d8a5ccbac2f/kube-rbac-proxy/0.log" Oct 01 17:01:56 crc kubenswrapper[4726]: I1001 17:01:56.777474 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-cpbbq_8e085c54-533c-4186-a6fe-5d8a5ccbac2f/manager/0.log" Oct 01 17:01:56 crc kubenswrapper[4726]: I1001 17:01:56.890714 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-5hvl7_23922646-bd87-424b-85d7-8a2a55056438/kube-rbac-proxy/0.log" Oct 01 17:01:56 crc kubenswrapper[4726]: I1001 17:01:56.988317 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-5hvl7_23922646-bd87-424b-85d7-8a2a55056438/manager/0.log" Oct 01 17:01:57 crc kubenswrapper[4726]: I1001 17:01:57.086923 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-qd8pc_2d90550d-aedc-4e50-81f9-da6f285f8c2b/kube-rbac-proxy/0.log" Oct 01 17:01:57 crc kubenswrapper[4726]: I1001 17:01:57.155463 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-qd8pc_2d90550d-aedc-4e50-81f9-da6f285f8c2b/manager/0.log" Oct 01 17:01:57 crc kubenswrapper[4726]: I1001 17:01:57.210765 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-5vlmb_9f6777bb-5041-4c66-967a-069fa217bf99/kube-rbac-proxy/0.log" Oct 01 17:01:57 crc kubenswrapper[4726]: I1001 17:01:57.316642 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-5vlmb_9f6777bb-5041-4c66-967a-069fa217bf99/manager/0.log" Oct 01 17:01:57 crc kubenswrapper[4726]: I1001 17:01:57.429856 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-w9xkn_0d89451f-e13a-4195-a2b2-f16443c43463/kube-rbac-proxy/0.log" Oct 01 17:01:57 crc kubenswrapper[4726]: I1001 17:01:57.477286 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-w9xkn_0d89451f-e13a-4195-a2b2-f16443c43463/manager/0.log" Oct 01 17:01:58 crc kubenswrapper[4726]: I1001 17:01:58.107026 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-zdkvq_0c8fbaaf-4ce0-4388-855e-fb85534f05e6/kube-rbac-proxy/0.log" Oct 01 17:01:58 crc kubenswrapper[4726]: I1001 17:01:58.298323 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-t2vwz_b9f1f46c-c6fa-443f-b53d-54614779c0cc/kube-rbac-proxy/0.log" Oct 01 17:01:58 crc kubenswrapper[4726]: I1001 17:01:58.323562 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-zdkvq_0c8fbaaf-4ce0-4388-855e-fb85534f05e6/manager/0.log" Oct 01 17:01:58 crc kubenswrapper[4726]: I1001 17:01:58.355588 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-t2vwz_b9f1f46c-c6fa-443f-b53d-54614779c0cc/manager/0.log" Oct 01 17:01:58 crc kubenswrapper[4726]: I1001 17:01:58.473650 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf_420c55b8-4442-41fa-98a4-338f7338ceb4/kube-rbac-proxy/0.log" Oct 01 17:01:58 crc kubenswrapper[4726]: I1001 17:01:58.504675 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf_420c55b8-4442-41fa-98a4-338f7338ceb4/manager/0.log" Oct 01 17:01:58 crc kubenswrapper[4726]: I1001 17:01:58.673266 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5f569c4895-tl4ld_42f6c148-bcc5-4054-88da-dc79d49baeff/kube-rbac-proxy/0.log" Oct 01 17:01:58 crc kubenswrapper[4726]: I1001 17:01:58.730667 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6877df559-w8rr9_07c756cd-913e-4237-a8eb-06dde60b6d4b/kube-rbac-proxy/0.log" Oct 01 17:01:58 crc kubenswrapper[4726]: I1001 17:01:58.909697 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-wh8qd_62b6b99f-a44a-445b-9183-917d39934259/registry-server/0.log" Oct 01 17:01:58 crc kubenswrapper[4726]: I1001 17:01:58.931850 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6877df559-w8rr9_07c756cd-913e-4237-a8eb-06dde60b6d4b/operator/0.log" Oct 01 17:01:59 crc kubenswrapper[4726]: I1001 17:01:59.121705 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-j4t5g_1845e7cd-8784-4994-8f4f-3015bf7d6b3b/kube-rbac-proxy/0.log" Oct 01 17:01:59 crc kubenswrapper[4726]: I1001 17:01:59.196216 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-j4t5g_1845e7cd-8784-4994-8f4f-3015bf7d6b3b/manager/0.log" Oct 01 17:01:59 crc kubenswrapper[4726]: I1001 17:01:59.657559 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-7s6mh_ff753ea2-fa5b-4367-b476-596fd49d0557/kube-rbac-proxy/0.log" Oct 01 17:01:59 crc kubenswrapper[4726]: I1001 17:01:59.724843 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5f569c4895-tl4ld_42f6c148-bcc5-4054-88da-dc79d49baeff/manager/0.log" Oct 01 17:01:59 crc kubenswrapper[4726]: I1001 17:01:59.808610 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:01:59 crc kubenswrapper[4726]: E1001 17:01:59.808825 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:01:59 crc kubenswrapper[4726]: I1001 17:01:59.898322 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv_965ea363-f11c-4ced-9988-debe220ee3f9/operator/0.log" Oct 01 17:01:59 crc kubenswrapper[4726]: I1001 17:01:59.907015 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-pl7jv_fa99155f-85e2-406e-84c4-4a95ab74407f/kube-rbac-proxy/0.log" Oct 01 17:01:59 crc kubenswrapper[4726]: I1001 17:01:59.956778 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-7s6mh_ff753ea2-fa5b-4367-b476-596fd49d0557/manager/0.log" Oct 01 17:02:00 crc kubenswrapper[4726]: I1001 17:02:00.112776 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-pl7jv_fa99155f-85e2-406e-84c4-4a95ab74407f/manager/0.log" Oct 01 17:02:00 crc kubenswrapper[4726]: I1001 17:02:00.120981 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7866c8b47c-z6rlg_75c9f506-beb2-4fe6-9ada-3ed684131102/kube-rbac-proxy/0.log" Oct 01 17:02:00 crc kubenswrapper[4726]: I1001 17:02:00.169114 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7866c8b47c-z6rlg_75c9f506-beb2-4fe6-9ada-3ed684131102/manager/0.log" Oct 01 17:02:00 crc kubenswrapper[4726]: I1001 17:02:00.329148 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-9nqm2_b824be4d-8e37-42b5-8fe2-0dc6a1c4b4a1/kube-rbac-proxy/0.log" Oct 01 17:02:00 crc kubenswrapper[4726]: I1001 17:02:00.334460 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-nv4ww_0396f184-2150-4944-a81c-c0d0fafe1317/manager/0.log" Oct 01 17:02:00 crc kubenswrapper[4726]: I1001 17:02:00.382703 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-9nqm2_b824be4d-8e37-42b5-8fe2-0dc6a1c4b4a1/manager/0.log" Oct 01 17:02:00 crc kubenswrapper[4726]: I1001 17:02:00.383890 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-nv4ww_0396f184-2150-4944-a81c-c0d0fafe1317/kube-rbac-proxy/0.log" Oct 01 17:02:10 crc kubenswrapper[4726]: I1001 17:02:10.808445 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:02:10 crc kubenswrapper[4726]: E1001 17:02:10.809083 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:02:16 crc kubenswrapper[4726]: I1001 17:02:16.844081 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-z4ml2_8282f47b-cd59-4cba-a1aa-7528d57842ec/control-plane-machine-set-operator/0.log" Oct 01 17:02:16 crc kubenswrapper[4726]: I1001 17:02:16.970733 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mpfpx_5399581d-2111-47d1-bfe7-007b7c46d706/kube-rbac-proxy/0.log" Oct 01 17:02:17 crc kubenswrapper[4726]: I1001 17:02:17.042920 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mpfpx_5399581d-2111-47d1-bfe7-007b7c46d706/machine-api-operator/0.log" Oct 01 17:02:25 crc kubenswrapper[4726]: I1001 17:02:25.808366 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:02:25 crc kubenswrapper[4726]: E1001 17:02:25.809462 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:02:29 crc kubenswrapper[4726]: I1001 17:02:29.484248 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-jkfpv_cfc89425-6bca-4267-9a51-50e0d8325eaa/cert-manager-controller/0.log" Oct 01 17:02:29 crc kubenswrapper[4726]: I1001 17:02:29.654585 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-pz8pb_55df3602-b3df-45f4-90b8-eddb1903bf32/cert-manager-cainjector/0.log" Oct 01 17:02:29 crc kubenswrapper[4726]: I1001 17:02:29.707121 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-jxd7m_51706d0c-886c-4865-ba8b-ade37f39b84e/cert-manager-webhook/0.log" Oct 01 17:02:38 crc kubenswrapper[4726]: I1001 17:02:38.821387 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:02:38 crc kubenswrapper[4726]: E1001 17:02:38.822369 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:02:41 crc kubenswrapper[4726]: I1001 17:02:41.603167 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-vqzfm_5fb896f2-8fbc-40d7-adf4-2a93089171ce/nmstate-console-plugin/0.log" Oct 01 17:02:41 crc kubenswrapper[4726]: I1001 17:02:41.755853 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-kzwz4_fc59e6f9-76eb-49fb-9028-9e64820de406/nmstate-handler/0.log" Oct 01 17:02:41 crc kubenswrapper[4726]: I1001 17:02:41.770954 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-xxpvz_d24b198c-db7b-4057-bed1-6f32d0bac5d3/kube-rbac-proxy/0.log" Oct 01 17:02:41 crc kubenswrapper[4726]: I1001 17:02:41.784423 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-xxpvz_d24b198c-db7b-4057-bed1-6f32d0bac5d3/nmstate-metrics/0.log" Oct 01 17:02:41 crc kubenswrapper[4726]: I1001 17:02:41.931926 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-mn7jt_d4e4c724-1025-48b6-a04c-77071be69269/nmstate-operator/0.log" Oct 01 17:02:41 crc kubenswrapper[4726]: I1001 17:02:41.975959 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-qrkz2_0f80c25b-98a7-46d0-bf07-26c19f885781/nmstate-webhook/0.log" Oct 01 17:02:49 crc kubenswrapper[4726]: I1001 17:02:49.808269 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:02:49 crc kubenswrapper[4726]: E1001 17:02:49.808967 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:02:55 crc kubenswrapper[4726]: I1001 17:02:55.972065 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-kk7mh_1c256ab6-be12-4843-a072-f6bb861c9740/kube-rbac-proxy/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.069365 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-kk7mh_1c256ab6-be12-4843-a072-f6bb861c9740/controller/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.262162 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-mqdn2_678af4b7-7942-40da-8273-de583fd22666/frr-k8s-webhook-server/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.312369 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-frr-files/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.461230 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-frr-files/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.474788 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-reloader/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.481102 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-metrics/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.542878 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-reloader/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.649965 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-frr-files/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.692768 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-reloader/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.694017 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-metrics/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.763839 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-metrics/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.927476 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-frr-files/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.944503 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-reloader/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.970206 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-metrics/0.log" Oct 01 17:02:56 crc kubenswrapper[4726]: I1001 17:02:56.997713 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/controller/0.log" Oct 01 17:02:57 crc kubenswrapper[4726]: I1001 17:02:57.179088 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/frr-metrics/0.log" Oct 01 17:02:57 crc kubenswrapper[4726]: I1001 17:02:57.203504 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/kube-rbac-proxy/0.log" Oct 01 17:02:57 crc kubenswrapper[4726]: I1001 17:02:57.243579 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/kube-rbac-proxy-frr/0.log" Oct 01 17:02:57 crc kubenswrapper[4726]: I1001 17:02:57.414474 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/reloader/0.log" Oct 01 17:02:57 crc kubenswrapper[4726]: I1001 17:02:57.451740 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/frr/0.log" Oct 01 17:02:57 crc kubenswrapper[4726]: I1001 17:02:57.471003 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5fc7dbb978-2m6rr_9dd491a2-a155-401a-86fb-04e5fe23984e/manager/0.log" Oct 01 17:02:57 crc kubenswrapper[4726]: I1001 17:02:57.627337 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-d8cc586f4-l28zk_f729348c-0bd7-4667-86f9-94e36a9afb11/webhook-server/0.log" Oct 01 17:02:57 crc kubenswrapper[4726]: I1001 17:02:57.653114 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tdc4_087620f7-11b5-49b6-a8b7-30a75a5196cb/kube-rbac-proxy/0.log" Oct 01 17:02:57 crc kubenswrapper[4726]: I1001 17:02:57.954215 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tdc4_087620f7-11b5-49b6-a8b7-30a75a5196cb/speaker/0.log" Oct 01 17:03:02 crc kubenswrapper[4726]: I1001 17:03:02.808741 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:03:02 crc kubenswrapper[4726]: E1001 17:03:02.809871 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.326204 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7f82c"] Oct 01 17:03:08 crc kubenswrapper[4726]: E1001 17:03:08.327342 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="170e63ef-d054-4975-a6c9-6ec4b114ca58" containerName="container-00" Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.327359 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="170e63ef-d054-4975-a6c9-6ec4b114ca58" containerName="container-00" Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.327563 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="170e63ef-d054-4975-a6c9-6ec4b114ca58" containerName="container-00" Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.329222 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.337188 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7f82c"] Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.480670 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhk52\" (UniqueName: \"kubernetes.io/projected/fe3e5196-9359-442f-8b88-0d9afcfc018f-kube-api-access-rhk52\") pod \"redhat-operators-7f82c\" (UID: \"fe3e5196-9359-442f-8b88-0d9afcfc018f\") " pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.480777 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe3e5196-9359-442f-8b88-0d9afcfc018f-catalog-content\") pod \"redhat-operators-7f82c\" (UID: \"fe3e5196-9359-442f-8b88-0d9afcfc018f\") " pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.480810 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe3e5196-9359-442f-8b88-0d9afcfc018f-utilities\") pod \"redhat-operators-7f82c\" (UID: \"fe3e5196-9359-442f-8b88-0d9afcfc018f\") " pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.582333 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhk52\" (UniqueName: \"kubernetes.io/projected/fe3e5196-9359-442f-8b88-0d9afcfc018f-kube-api-access-rhk52\") pod \"redhat-operators-7f82c\" (UID: \"fe3e5196-9359-442f-8b88-0d9afcfc018f\") " pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.582408 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe3e5196-9359-442f-8b88-0d9afcfc018f-catalog-content\") pod \"redhat-operators-7f82c\" (UID: \"fe3e5196-9359-442f-8b88-0d9afcfc018f\") " pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.582430 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe3e5196-9359-442f-8b88-0d9afcfc018f-utilities\") pod \"redhat-operators-7f82c\" (UID: \"fe3e5196-9359-442f-8b88-0d9afcfc018f\") " pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.582915 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe3e5196-9359-442f-8b88-0d9afcfc018f-utilities\") pod \"redhat-operators-7f82c\" (UID: \"fe3e5196-9359-442f-8b88-0d9afcfc018f\") " pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.583690 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe3e5196-9359-442f-8b88-0d9afcfc018f-catalog-content\") pod \"redhat-operators-7f82c\" (UID: \"fe3e5196-9359-442f-8b88-0d9afcfc018f\") " pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.602809 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhk52\" (UniqueName: \"kubernetes.io/projected/fe3e5196-9359-442f-8b88-0d9afcfc018f-kube-api-access-rhk52\") pod \"redhat-operators-7f82c\" (UID: \"fe3e5196-9359-442f-8b88-0d9afcfc018f\") " pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:08 crc kubenswrapper[4726]: I1001 17:03:08.652036 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:09 crc kubenswrapper[4726]: I1001 17:03:09.127585 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7f82c"] Oct 01 17:03:10 crc kubenswrapper[4726]: I1001 17:03:10.117672 4726 generic.go:334] "Generic (PLEG): container finished" podID="fe3e5196-9359-442f-8b88-0d9afcfc018f" containerID="cc97ee47dfcedca58bdd09ec3dfd591f517b41ae60fe61a02e96ee055ab0f3df" exitCode=0 Oct 01 17:03:10 crc kubenswrapper[4726]: I1001 17:03:10.117955 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f82c" event={"ID":"fe3e5196-9359-442f-8b88-0d9afcfc018f","Type":"ContainerDied","Data":"cc97ee47dfcedca58bdd09ec3dfd591f517b41ae60fe61a02e96ee055ab0f3df"} Oct 01 17:03:10 crc kubenswrapper[4726]: I1001 17:03:10.117982 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f82c" event={"ID":"fe3e5196-9359-442f-8b88-0d9afcfc018f","Type":"ContainerStarted","Data":"b2ec3201ae75033c75b792260d5354918aaf602f343ee216c9fbabfb48d04c1d"} Oct 01 17:03:10 crc kubenswrapper[4726]: I1001 17:03:10.119891 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 17:03:11 crc kubenswrapper[4726]: I1001 17:03:11.321294 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2_7b3a32c8-9c47-4ccc-9885-d29009a22aa4/util/0.log" Oct 01 17:03:11 crc kubenswrapper[4726]: I1001 17:03:11.482462 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2_7b3a32c8-9c47-4ccc-9885-d29009a22aa4/util/0.log" Oct 01 17:03:11 crc kubenswrapper[4726]: I1001 17:03:11.490700 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2_7b3a32c8-9c47-4ccc-9885-d29009a22aa4/pull/0.log" Oct 01 17:03:11 crc kubenswrapper[4726]: I1001 17:03:11.524863 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2_7b3a32c8-9c47-4ccc-9885-d29009a22aa4/pull/0.log" Oct 01 17:03:11 crc kubenswrapper[4726]: I1001 17:03:11.703205 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2_7b3a32c8-9c47-4ccc-9885-d29009a22aa4/util/0.log" Oct 01 17:03:11 crc kubenswrapper[4726]: I1001 17:03:11.737667 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2_7b3a32c8-9c47-4ccc-9885-d29009a22aa4/extract/0.log" Oct 01 17:03:11 crc kubenswrapper[4726]: I1001 17:03:11.738710 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2_7b3a32c8-9c47-4ccc-9885-d29009a22aa4/pull/0.log" Oct 01 17:03:11 crc kubenswrapper[4726]: I1001 17:03:11.897694 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_f2e6c6e1-9224-4c26-adcf-07fdf1ea4538/util/0.log" Oct 01 17:03:12 crc kubenswrapper[4726]: I1001 17:03:12.030778 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_f2e6c6e1-9224-4c26-adcf-07fdf1ea4538/pull/0.log" Oct 01 17:03:12 crc kubenswrapper[4726]: I1001 17:03:12.072024 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_f2e6c6e1-9224-4c26-adcf-07fdf1ea4538/pull/0.log" Oct 01 17:03:12 crc kubenswrapper[4726]: I1001 17:03:12.075952 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_f2e6c6e1-9224-4c26-adcf-07fdf1ea4538/util/0.log" Oct 01 17:03:12 crc kubenswrapper[4726]: I1001 17:03:12.161879 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f82c" event={"ID":"fe3e5196-9359-442f-8b88-0d9afcfc018f","Type":"ContainerStarted","Data":"f4dc498758b1860d5b1149f025950f8ece0663b1536add46dcc7ed295a7515c7"} Oct 01 17:03:12 crc kubenswrapper[4726]: I1001 17:03:12.328332 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_f2e6c6e1-9224-4c26-adcf-07fdf1ea4538/extract/0.log" Oct 01 17:03:12 crc kubenswrapper[4726]: I1001 17:03:12.389770 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_f2e6c6e1-9224-4c26-adcf-07fdf1ea4538/pull/0.log" Oct 01 17:03:12 crc kubenswrapper[4726]: I1001 17:03:12.415669 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_f2e6c6e1-9224-4c26-adcf-07fdf1ea4538/util/0.log" Oct 01 17:03:12 crc kubenswrapper[4726]: I1001 17:03:12.516791 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q5bzx_9f2ddc53-446b-42b0-966a-f6b2f4eb6159/extract-utilities/0.log" Oct 01 17:03:12 crc kubenswrapper[4726]: I1001 17:03:12.686183 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q5bzx_9f2ddc53-446b-42b0-966a-f6b2f4eb6159/extract-content/0.log" Oct 01 17:03:12 crc kubenswrapper[4726]: I1001 17:03:12.702965 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q5bzx_9f2ddc53-446b-42b0-966a-f6b2f4eb6159/extract-utilities/0.log" Oct 01 17:03:12 crc kubenswrapper[4726]: I1001 17:03:12.726606 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q5bzx_9f2ddc53-446b-42b0-966a-f6b2f4eb6159/extract-content/0.log" Oct 01 17:03:12 crc kubenswrapper[4726]: I1001 17:03:12.973117 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q5bzx_9f2ddc53-446b-42b0-966a-f6b2f4eb6159/extract-utilities/0.log" Oct 01 17:03:12 crc kubenswrapper[4726]: I1001 17:03:12.995630 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q5bzx_9f2ddc53-446b-42b0-966a-f6b2f4eb6159/extract-content/0.log" Oct 01 17:03:13 crc kubenswrapper[4726]: I1001 17:03:13.176034 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ch575_039da856-2b47-4aff-b18b-9e9cff090200/extract-utilities/0.log" Oct 01 17:03:13 crc kubenswrapper[4726]: I1001 17:03:13.418703 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q5bzx_9f2ddc53-446b-42b0-966a-f6b2f4eb6159/registry-server/0.log" Oct 01 17:03:13 crc kubenswrapper[4726]: I1001 17:03:13.451605 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ch575_039da856-2b47-4aff-b18b-9e9cff090200/extract-content/0.log" Oct 01 17:03:13 crc kubenswrapper[4726]: I1001 17:03:13.451751 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ch575_039da856-2b47-4aff-b18b-9e9cff090200/extract-utilities/0.log" Oct 01 17:03:13 crc kubenswrapper[4726]: I1001 17:03:13.495930 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ch575_039da856-2b47-4aff-b18b-9e9cff090200/extract-content/0.log" Oct 01 17:03:13 crc kubenswrapper[4726]: I1001 17:03:13.718621 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ch575_039da856-2b47-4aff-b18b-9e9cff090200/extract-utilities/0.log" Oct 01 17:03:13 crc kubenswrapper[4726]: I1001 17:03:13.737948 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ch575_039da856-2b47-4aff-b18b-9e9cff090200/extract-content/0.log" Oct 01 17:03:13 crc kubenswrapper[4726]: I1001 17:03:13.928527 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b_b751f5f8-ea87-4fcb-8c07-93863ae71d03/util/0.log" Oct 01 17:03:14 crc kubenswrapper[4726]: I1001 17:03:14.209180 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b_b751f5f8-ea87-4fcb-8c07-93863ae71d03/pull/0.log" Oct 01 17:03:14 crc kubenswrapper[4726]: I1001 17:03:14.242935 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b_b751f5f8-ea87-4fcb-8c07-93863ae71d03/pull/0.log" Oct 01 17:03:14 crc kubenswrapper[4726]: I1001 17:03:14.504263 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ch575_039da856-2b47-4aff-b18b-9e9cff090200/registry-server/0.log" Oct 01 17:03:14 crc kubenswrapper[4726]: I1001 17:03:14.702367 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b_b751f5f8-ea87-4fcb-8c07-93863ae71d03/extract/0.log" Oct 01 17:03:14 crc kubenswrapper[4726]: I1001 17:03:14.879155 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b_b751f5f8-ea87-4fcb-8c07-93863ae71d03/util/0.log" Oct 01 17:03:14 crc kubenswrapper[4726]: I1001 17:03:14.929894 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b_b751f5f8-ea87-4fcb-8c07-93863ae71d03/util/0.log" Oct 01 17:03:15 crc kubenswrapper[4726]: I1001 17:03:15.135023 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b_b751f5f8-ea87-4fcb-8c07-93863ae71d03/pull/0.log" Oct 01 17:03:15 crc kubenswrapper[4726]: I1001 17:03:15.321028 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s_213c0beb-44a8-4e18-afaa-4d2ba6fd6305/util/0.log" Oct 01 17:03:15 crc kubenswrapper[4726]: I1001 17:03:15.321184 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s_213c0beb-44a8-4e18-afaa-4d2ba6fd6305/util/0.log" Oct 01 17:03:15 crc kubenswrapper[4726]: I1001 17:03:15.322196 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s_213c0beb-44a8-4e18-afaa-4d2ba6fd6305/pull/0.log" Oct 01 17:03:15 crc kubenswrapper[4726]: I1001 17:03:15.322562 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s_213c0beb-44a8-4e18-afaa-4d2ba6fd6305/pull/0.log" Oct 01 17:03:15 crc kubenswrapper[4726]: I1001 17:03:15.502882 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s_213c0beb-44a8-4e18-afaa-4d2ba6fd6305/util/0.log" Oct 01 17:03:15 crc kubenswrapper[4726]: I1001 17:03:15.518826 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s_213c0beb-44a8-4e18-afaa-4d2ba6fd6305/pull/0.log" Oct 01 17:03:15 crc kubenswrapper[4726]: I1001 17:03:15.574009 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-g69q5_c460d271-b44c-482b-a0d9-70bdc879cf1d/marketplace-operator/0.log" Oct 01 17:03:15 crc kubenswrapper[4726]: I1001 17:03:15.576302 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s_213c0beb-44a8-4e18-afaa-4d2ba6fd6305/extract/0.log" Oct 01 17:03:15 crc kubenswrapper[4726]: I1001 17:03:15.917002 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ms6sz_4753c482-c4a6-44e9-ace2-7997ee15d241/extract-utilities/0.log" Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.088000 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ms6sz_4753c482-c4a6-44e9-ace2-7997ee15d241/extract-content/0.log" Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.120790 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ms6sz_4753c482-c4a6-44e9-ace2-7997ee15d241/extract-content/0.log" Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.126896 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ms6sz_4753c482-c4a6-44e9-ace2-7997ee15d241/extract-utilities/0.log" Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.197227 4726 generic.go:334] "Generic (PLEG): container finished" podID="fe3e5196-9359-442f-8b88-0d9afcfc018f" containerID="f4dc498758b1860d5b1149f025950f8ece0663b1536add46dcc7ed295a7515c7" exitCode=0 Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.197278 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f82c" event={"ID":"fe3e5196-9359-442f-8b88-0d9afcfc018f","Type":"ContainerDied","Data":"f4dc498758b1860d5b1149f025950f8ece0663b1536add46dcc7ed295a7515c7"} Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.267351 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ms6sz_4753c482-c4a6-44e9-ace2-7997ee15d241/extract-content/0.log" Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.285469 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ms6sz_4753c482-c4a6-44e9-ace2-7997ee15d241/extract-utilities/0.log" Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.357291 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7f82c_fe3e5196-9359-442f-8b88-0d9afcfc018f/extract-utilities/0.log" Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.547568 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ms6sz_4753c482-c4a6-44e9-ace2-7997ee15d241/registry-server/0.log" Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.613028 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7f82c_fe3e5196-9359-442f-8b88-0d9afcfc018f/extract-content/0.log" Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.613069 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7f82c_fe3e5196-9359-442f-8b88-0d9afcfc018f/extract-utilities/0.log" Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.641308 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7f82c_fe3e5196-9359-442f-8b88-0d9afcfc018f/extract-content/0.log" Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.773132 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7f82c_fe3e5196-9359-442f-8b88-0d9afcfc018f/extract-utilities/0.log" Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.828535 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7f82c_fe3e5196-9359-442f-8b88-0d9afcfc018f/extract-content/0.log" Oct 01 17:03:16 crc kubenswrapper[4726]: I1001 17:03:16.866149 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nj4wf_c24a10a9-6d60-43d5-9e3c-67cec3e0045c/extract-utilities/0.log" Oct 01 17:03:17 crc kubenswrapper[4726]: I1001 17:03:17.000611 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nj4wf_c24a10a9-6d60-43d5-9e3c-67cec3e0045c/extract-content/0.log" Oct 01 17:03:17 crc kubenswrapper[4726]: I1001 17:03:17.004090 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nj4wf_c24a10a9-6d60-43d5-9e3c-67cec3e0045c/extract-utilities/0.log" Oct 01 17:03:17 crc kubenswrapper[4726]: I1001 17:03:17.055233 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nj4wf_c24a10a9-6d60-43d5-9e3c-67cec3e0045c/extract-content/0.log" Oct 01 17:03:17 crc kubenswrapper[4726]: I1001 17:03:17.206467 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f82c" event={"ID":"fe3e5196-9359-442f-8b88-0d9afcfc018f","Type":"ContainerStarted","Data":"9f550cc4eca532eb5554a6cc49b298744885a461d3fb92fdd15720bd05f1e87a"} Oct 01 17:03:17 crc kubenswrapper[4726]: I1001 17:03:17.226835 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7f82c" podStartSLOduration=2.472831673 podStartE2EDuration="9.226817747s" podCreationTimestamp="2025-10-01 17:03:08 +0000 UTC" firstStartedPulling="2025-10-01 17:03:10.119350568 +0000 UTC m=+3843.020903145" lastFinishedPulling="2025-10-01 17:03:16.873336652 +0000 UTC m=+3849.774889219" observedRunningTime="2025-10-01 17:03:17.225249792 +0000 UTC m=+3850.126802369" watchObservedRunningTime="2025-10-01 17:03:17.226817747 +0000 UTC m=+3850.128370324" Oct 01 17:03:17 crc kubenswrapper[4726]: I1001 17:03:17.262090 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nj4wf_c24a10a9-6d60-43d5-9e3c-67cec3e0045c/extract-content/0.log" Oct 01 17:03:17 crc kubenswrapper[4726]: I1001 17:03:17.318416 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nj4wf_c24a10a9-6d60-43d5-9e3c-67cec3e0045c/extract-utilities/0.log" Oct 01 17:03:17 crc kubenswrapper[4726]: I1001 17:03:17.571551 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nj4wf_c24a10a9-6d60-43d5-9e3c-67cec3e0045c/registry-server/0.log" Oct 01 17:03:17 crc kubenswrapper[4726]: I1001 17:03:17.813894 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:03:17 crc kubenswrapper[4726]: E1001 17:03:17.814398 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:03:18 crc kubenswrapper[4726]: I1001 17:03:18.652566 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:18 crc kubenswrapper[4726]: I1001 17:03:18.653800 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:19 crc kubenswrapper[4726]: I1001 17:03:19.705153 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7f82c" podUID="fe3e5196-9359-442f-8b88-0d9afcfc018f" containerName="registry-server" probeResult="failure" output=< Oct 01 17:03:19 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Oct 01 17:03:19 crc kubenswrapper[4726]: > Oct 01 17:03:28 crc kubenswrapper[4726]: I1001 17:03:28.808637 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:03:28 crc kubenswrapper[4726]: E1001 17:03:28.809572 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:03:28 crc kubenswrapper[4726]: I1001 17:03:28.985922 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:29 crc kubenswrapper[4726]: I1001 17:03:29.064431 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:29 crc kubenswrapper[4726]: I1001 17:03:29.226381 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7f82c"] Oct 01 17:03:30 crc kubenswrapper[4726]: I1001 17:03:30.332545 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7f82c" podUID="fe3e5196-9359-442f-8b88-0d9afcfc018f" containerName="registry-server" containerID="cri-o://9f550cc4eca532eb5554a6cc49b298744885a461d3fb92fdd15720bd05f1e87a" gracePeriod=2 Oct 01 17:03:31 crc kubenswrapper[4726]: I1001 17:03:31.345692 4726 generic.go:334] "Generic (PLEG): container finished" podID="fe3e5196-9359-442f-8b88-0d9afcfc018f" containerID="9f550cc4eca532eb5554a6cc49b298744885a461d3fb92fdd15720bd05f1e87a" exitCode=0 Oct 01 17:03:31 crc kubenswrapper[4726]: I1001 17:03:31.345908 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f82c" event={"ID":"fe3e5196-9359-442f-8b88-0d9afcfc018f","Type":"ContainerDied","Data":"9f550cc4eca532eb5554a6cc49b298744885a461d3fb92fdd15720bd05f1e87a"} Oct 01 17:03:31 crc kubenswrapper[4726]: I1001 17:03:31.346148 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f82c" event={"ID":"fe3e5196-9359-442f-8b88-0d9afcfc018f","Type":"ContainerDied","Data":"b2ec3201ae75033c75b792260d5354918aaf602f343ee216c9fbabfb48d04c1d"} Oct 01 17:03:31 crc kubenswrapper[4726]: I1001 17:03:31.346161 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2ec3201ae75033c75b792260d5354918aaf602f343ee216c9fbabfb48d04c1d" Oct 01 17:03:31 crc kubenswrapper[4726]: I1001 17:03:31.429825 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:31 crc kubenswrapper[4726]: I1001 17:03:31.501201 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe3e5196-9359-442f-8b88-0d9afcfc018f-catalog-content\") pod \"fe3e5196-9359-442f-8b88-0d9afcfc018f\" (UID: \"fe3e5196-9359-442f-8b88-0d9afcfc018f\") " Oct 01 17:03:31 crc kubenswrapper[4726]: I1001 17:03:31.501247 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhk52\" (UniqueName: \"kubernetes.io/projected/fe3e5196-9359-442f-8b88-0d9afcfc018f-kube-api-access-rhk52\") pod \"fe3e5196-9359-442f-8b88-0d9afcfc018f\" (UID: \"fe3e5196-9359-442f-8b88-0d9afcfc018f\") " Oct 01 17:03:31 crc kubenswrapper[4726]: I1001 17:03:31.501290 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe3e5196-9359-442f-8b88-0d9afcfc018f-utilities\") pod \"fe3e5196-9359-442f-8b88-0d9afcfc018f\" (UID: \"fe3e5196-9359-442f-8b88-0d9afcfc018f\") " Oct 01 17:03:31 crc kubenswrapper[4726]: I1001 17:03:31.502189 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe3e5196-9359-442f-8b88-0d9afcfc018f-utilities" (OuterVolumeSpecName: "utilities") pod "fe3e5196-9359-442f-8b88-0d9afcfc018f" (UID: "fe3e5196-9359-442f-8b88-0d9afcfc018f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 17:03:31 crc kubenswrapper[4726]: I1001 17:03:31.507770 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe3e5196-9359-442f-8b88-0d9afcfc018f-kube-api-access-rhk52" (OuterVolumeSpecName: "kube-api-access-rhk52") pod "fe3e5196-9359-442f-8b88-0d9afcfc018f" (UID: "fe3e5196-9359-442f-8b88-0d9afcfc018f"). InnerVolumeSpecName "kube-api-access-rhk52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 17:03:31 crc kubenswrapper[4726]: I1001 17:03:31.588303 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe3e5196-9359-442f-8b88-0d9afcfc018f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe3e5196-9359-442f-8b88-0d9afcfc018f" (UID: "fe3e5196-9359-442f-8b88-0d9afcfc018f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 17:03:31 crc kubenswrapper[4726]: I1001 17:03:31.604232 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe3e5196-9359-442f-8b88-0d9afcfc018f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 17:03:31 crc kubenswrapper[4726]: I1001 17:03:31.604309 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhk52\" (UniqueName: \"kubernetes.io/projected/fe3e5196-9359-442f-8b88-0d9afcfc018f-kube-api-access-rhk52\") on node \"crc\" DevicePath \"\"" Oct 01 17:03:31 crc kubenswrapper[4726]: I1001 17:03:31.604334 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe3e5196-9359-442f-8b88-0d9afcfc018f-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 17:03:32 crc kubenswrapper[4726]: I1001 17:03:32.356801 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f82c" Oct 01 17:03:32 crc kubenswrapper[4726]: I1001 17:03:32.382030 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7f82c"] Oct 01 17:03:32 crc kubenswrapper[4726]: I1001 17:03:32.391655 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7f82c"] Oct 01 17:03:33 crc kubenswrapper[4726]: I1001 17:03:33.823165 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe3e5196-9359-442f-8b88-0d9afcfc018f" path="/var/lib/kubelet/pods/fe3e5196-9359-442f-8b88-0d9afcfc018f/volumes" Oct 01 17:03:40 crc kubenswrapper[4726]: I1001 17:03:40.808406 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:03:40 crc kubenswrapper[4726]: E1001 17:03:40.809275 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:03:46 crc kubenswrapper[4726]: E1001 17:03:46.363018 4726 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.217:60980->38.102.83.217:36913: write tcp 38.102.83.217:60980->38.102.83.217:36913: write: broken pipe Oct 01 17:03:51 crc kubenswrapper[4726]: I1001 17:03:51.808934 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:03:51 crc kubenswrapper[4726]: E1001 17:03:51.811012 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:04:02 crc kubenswrapper[4726]: I1001 17:04:02.810456 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:04:02 crc kubenswrapper[4726]: E1001 17:04:02.812082 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:04:17 crc kubenswrapper[4726]: I1001 17:04:17.819846 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:04:17 crc kubenswrapper[4726]: E1001 17:04:17.822361 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:04:30 crc kubenswrapper[4726]: I1001 17:04:30.808118 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:04:30 crc kubenswrapper[4726]: E1001 17:04:30.810281 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:04:42 crc kubenswrapper[4726]: I1001 17:04:42.809219 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:04:42 crc kubenswrapper[4726]: E1001 17:04:42.810460 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:04:54 crc kubenswrapper[4726]: I1001 17:04:54.807360 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:04:55 crc kubenswrapper[4726]: I1001 17:04:55.234994 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"6a39a92e4a63907159963d04757f77099dc1afa2ff471236819de46bcfbf3573"} Oct 01 17:05:17 crc kubenswrapper[4726]: I1001 17:05:17.481699 4726 generic.go:334] "Generic (PLEG): container finished" podID="bdd1d196-704d-4ed3-a919-b4f1ec4ce217" containerID="7eb6686dfcaa5d8708af187bad76e73c912908a67c0ab1c5c6c5652e2c78c90c" exitCode=0 Oct 01 17:05:17 crc kubenswrapper[4726]: I1001 17:05:17.482279 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgg2v/must-gather-c94h2" event={"ID":"bdd1d196-704d-4ed3-a919-b4f1ec4ce217","Type":"ContainerDied","Data":"7eb6686dfcaa5d8708af187bad76e73c912908a67c0ab1c5c6c5652e2c78c90c"} Oct 01 17:05:17 crc kubenswrapper[4726]: I1001 17:05:17.483269 4726 scope.go:117] "RemoveContainer" containerID="7eb6686dfcaa5d8708af187bad76e73c912908a67c0ab1c5c6c5652e2c78c90c" Oct 01 17:05:17 crc kubenswrapper[4726]: I1001 17:05:17.824403 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jgg2v_must-gather-c94h2_bdd1d196-704d-4ed3-a919-b4f1ec4ce217/gather/0.log" Oct 01 17:05:25 crc kubenswrapper[4726]: I1001 17:05:25.871070 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jgg2v/must-gather-c94h2"] Oct 01 17:05:25 crc kubenswrapper[4726]: I1001 17:05:25.871986 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-jgg2v/must-gather-c94h2" podUID="bdd1d196-704d-4ed3-a919-b4f1ec4ce217" containerName="copy" containerID="cri-o://f6cb3e8504569a12ba7d40e115fab94a701a0cc7bd2b6490362ea6a2b9311b2e" gracePeriod=2 Oct 01 17:05:25 crc kubenswrapper[4726]: I1001 17:05:25.880699 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jgg2v/must-gather-c94h2"] Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.339597 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jgg2v_must-gather-c94h2_bdd1d196-704d-4ed3-a919-b4f1ec4ce217/copy/0.log" Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.340267 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/must-gather-c94h2" Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.408146 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwkwf\" (UniqueName: \"kubernetes.io/projected/bdd1d196-704d-4ed3-a919-b4f1ec4ce217-kube-api-access-qwkwf\") pod \"bdd1d196-704d-4ed3-a919-b4f1ec4ce217\" (UID: \"bdd1d196-704d-4ed3-a919-b4f1ec4ce217\") " Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.408212 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bdd1d196-704d-4ed3-a919-b4f1ec4ce217-must-gather-output\") pod \"bdd1d196-704d-4ed3-a919-b4f1ec4ce217\" (UID: \"bdd1d196-704d-4ed3-a919-b4f1ec4ce217\") " Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.417484 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdd1d196-704d-4ed3-a919-b4f1ec4ce217-kube-api-access-qwkwf" (OuterVolumeSpecName: "kube-api-access-qwkwf") pod "bdd1d196-704d-4ed3-a919-b4f1ec4ce217" (UID: "bdd1d196-704d-4ed3-a919-b4f1ec4ce217"). InnerVolumeSpecName "kube-api-access-qwkwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.511027 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwkwf\" (UniqueName: \"kubernetes.io/projected/bdd1d196-704d-4ed3-a919-b4f1ec4ce217-kube-api-access-qwkwf\") on node \"crc\" DevicePath \"\"" Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.586862 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jgg2v_must-gather-c94h2_bdd1d196-704d-4ed3-a919-b4f1ec4ce217/copy/0.log" Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.587475 4726 generic.go:334] "Generic (PLEG): container finished" podID="bdd1d196-704d-4ed3-a919-b4f1ec4ce217" containerID="f6cb3e8504569a12ba7d40e115fab94a701a0cc7bd2b6490362ea6a2b9311b2e" exitCode=143 Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.587562 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgg2v/must-gather-c94h2" Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.587635 4726 scope.go:117] "RemoveContainer" containerID="f6cb3e8504569a12ba7d40e115fab94a701a0cc7bd2b6490362ea6a2b9311b2e" Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.596583 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdd1d196-704d-4ed3-a919-b4f1ec4ce217-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "bdd1d196-704d-4ed3-a919-b4f1ec4ce217" (UID: "bdd1d196-704d-4ed3-a919-b4f1ec4ce217"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.613804 4726 scope.go:117] "RemoveContainer" containerID="7eb6686dfcaa5d8708af187bad76e73c912908a67c0ab1c5c6c5652e2c78c90c" Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.613879 4726 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bdd1d196-704d-4ed3-a919-b4f1ec4ce217-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.670038 4726 scope.go:117] "RemoveContainer" containerID="f6cb3e8504569a12ba7d40e115fab94a701a0cc7bd2b6490362ea6a2b9311b2e" Oct 01 17:05:26 crc kubenswrapper[4726]: E1001 17:05:26.670762 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6cb3e8504569a12ba7d40e115fab94a701a0cc7bd2b6490362ea6a2b9311b2e\": container with ID starting with f6cb3e8504569a12ba7d40e115fab94a701a0cc7bd2b6490362ea6a2b9311b2e not found: ID does not exist" containerID="f6cb3e8504569a12ba7d40e115fab94a701a0cc7bd2b6490362ea6a2b9311b2e" Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.670812 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6cb3e8504569a12ba7d40e115fab94a701a0cc7bd2b6490362ea6a2b9311b2e"} err="failed to get container status \"f6cb3e8504569a12ba7d40e115fab94a701a0cc7bd2b6490362ea6a2b9311b2e\": rpc error: code = NotFound desc = could not find container \"f6cb3e8504569a12ba7d40e115fab94a701a0cc7bd2b6490362ea6a2b9311b2e\": container with ID starting with f6cb3e8504569a12ba7d40e115fab94a701a0cc7bd2b6490362ea6a2b9311b2e not found: ID does not exist" Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.670842 4726 scope.go:117] "RemoveContainer" containerID="7eb6686dfcaa5d8708af187bad76e73c912908a67c0ab1c5c6c5652e2c78c90c" Oct 01 17:05:26 crc kubenswrapper[4726]: E1001 17:05:26.671353 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eb6686dfcaa5d8708af187bad76e73c912908a67c0ab1c5c6c5652e2c78c90c\": container with ID starting with 7eb6686dfcaa5d8708af187bad76e73c912908a67c0ab1c5c6c5652e2c78c90c not found: ID does not exist" containerID="7eb6686dfcaa5d8708af187bad76e73c912908a67c0ab1c5c6c5652e2c78c90c" Oct 01 17:05:26 crc kubenswrapper[4726]: I1001 17:05:26.671411 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eb6686dfcaa5d8708af187bad76e73c912908a67c0ab1c5c6c5652e2c78c90c"} err="failed to get container status \"7eb6686dfcaa5d8708af187bad76e73c912908a67c0ab1c5c6c5652e2c78c90c\": rpc error: code = NotFound desc = could not find container \"7eb6686dfcaa5d8708af187bad76e73c912908a67c0ab1c5c6c5652e2c78c90c\": container with ID starting with 7eb6686dfcaa5d8708af187bad76e73c912908a67c0ab1c5c6c5652e2c78c90c not found: ID does not exist" Oct 01 17:05:27 crc kubenswrapper[4726]: I1001 17:05:27.820670 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdd1d196-704d-4ed3-a919-b4f1ec4ce217" path="/var/lib/kubelet/pods/bdd1d196-704d-4ed3-a919-b4f1ec4ce217/volumes" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.460248 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lb9wc/must-gather-4pq84"] Oct 01 17:05:53 crc kubenswrapper[4726]: E1001 17:05:53.461877 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdd1d196-704d-4ed3-a919-b4f1ec4ce217" containerName="gather" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.461894 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdd1d196-704d-4ed3-a919-b4f1ec4ce217" containerName="gather" Oct 01 17:05:53 crc kubenswrapper[4726]: E1001 17:05:53.461910 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe3e5196-9359-442f-8b88-0d9afcfc018f" containerName="extract-content" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.461916 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe3e5196-9359-442f-8b88-0d9afcfc018f" containerName="extract-content" Oct 01 17:05:53 crc kubenswrapper[4726]: E1001 17:05:53.461928 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe3e5196-9359-442f-8b88-0d9afcfc018f" containerName="registry-server" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.461934 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe3e5196-9359-442f-8b88-0d9afcfc018f" containerName="registry-server" Oct 01 17:05:53 crc kubenswrapper[4726]: E1001 17:05:53.461953 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdd1d196-704d-4ed3-a919-b4f1ec4ce217" containerName="copy" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.461961 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdd1d196-704d-4ed3-a919-b4f1ec4ce217" containerName="copy" Oct 01 17:05:53 crc kubenswrapper[4726]: E1001 17:05:53.461980 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe3e5196-9359-442f-8b88-0d9afcfc018f" containerName="extract-utilities" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.461988 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe3e5196-9359-442f-8b88-0d9afcfc018f" containerName="extract-utilities" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.478893 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdd1d196-704d-4ed3-a919-b4f1ec4ce217" containerName="copy" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.479004 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdd1d196-704d-4ed3-a919-b4f1ec4ce217" containerName="gather" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.479031 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe3e5196-9359-442f-8b88-0d9afcfc018f" containerName="registry-server" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.480412 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-lb9wc/must-gather-4pq84"] Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.480526 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/must-gather-4pq84" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.488568 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-lb9wc"/"default-dockercfg-n6czv" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.488752 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-lb9wc"/"openshift-service-ca.crt" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.488836 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-lb9wc"/"kube-root-ca.crt" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.581951 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e49d5b8d-736e-440c-8cc1-66fd76db3937-must-gather-output\") pod \"must-gather-4pq84\" (UID: \"e49d5b8d-736e-440c-8cc1-66fd76db3937\") " pod="openshift-must-gather-lb9wc/must-gather-4pq84" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.582137 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlgvr\" (UniqueName: \"kubernetes.io/projected/e49d5b8d-736e-440c-8cc1-66fd76db3937-kube-api-access-zlgvr\") pod \"must-gather-4pq84\" (UID: \"e49d5b8d-736e-440c-8cc1-66fd76db3937\") " pod="openshift-must-gather-lb9wc/must-gather-4pq84" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.683779 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlgvr\" (UniqueName: \"kubernetes.io/projected/e49d5b8d-736e-440c-8cc1-66fd76db3937-kube-api-access-zlgvr\") pod \"must-gather-4pq84\" (UID: \"e49d5b8d-736e-440c-8cc1-66fd76db3937\") " pod="openshift-must-gather-lb9wc/must-gather-4pq84" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.683888 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e49d5b8d-736e-440c-8cc1-66fd76db3937-must-gather-output\") pod \"must-gather-4pq84\" (UID: \"e49d5b8d-736e-440c-8cc1-66fd76db3937\") " pod="openshift-must-gather-lb9wc/must-gather-4pq84" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.684428 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e49d5b8d-736e-440c-8cc1-66fd76db3937-must-gather-output\") pod \"must-gather-4pq84\" (UID: \"e49d5b8d-736e-440c-8cc1-66fd76db3937\") " pod="openshift-must-gather-lb9wc/must-gather-4pq84" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.703519 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlgvr\" (UniqueName: \"kubernetes.io/projected/e49d5b8d-736e-440c-8cc1-66fd76db3937-kube-api-access-zlgvr\") pod \"must-gather-4pq84\" (UID: \"e49d5b8d-736e-440c-8cc1-66fd76db3937\") " pod="openshift-must-gather-lb9wc/must-gather-4pq84" Oct 01 17:05:53 crc kubenswrapper[4726]: I1001 17:05:53.808023 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/must-gather-4pq84" Oct 01 17:05:54 crc kubenswrapper[4726]: I1001 17:05:54.264860 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-lb9wc/must-gather-4pq84"] Oct 01 17:05:54 crc kubenswrapper[4726]: I1001 17:05:54.901543 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lb9wc/must-gather-4pq84" event={"ID":"e49d5b8d-736e-440c-8cc1-66fd76db3937","Type":"ContainerStarted","Data":"d03642d7648d13172802e303100b6660ba590556a25bc8172cc922264c903a4c"} Oct 01 17:05:54 crc kubenswrapper[4726]: I1001 17:05:54.901850 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lb9wc/must-gather-4pq84" event={"ID":"e49d5b8d-736e-440c-8cc1-66fd76db3937","Type":"ContainerStarted","Data":"2494b9dd9c22b6c332a37075f5b395a57d0d3a59674fbf19258d31b4e8122230"} Oct 01 17:05:54 crc kubenswrapper[4726]: I1001 17:05:54.901866 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lb9wc/must-gather-4pq84" event={"ID":"e49d5b8d-736e-440c-8cc1-66fd76db3937","Type":"ContainerStarted","Data":"0307551b3e0c2e41f8ad26f95fd69e33b1466fa8c555815304df39a8c1eb69b0"} Oct 01 17:05:57 crc kubenswrapper[4726]: I1001 17:05:57.843603 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-lb9wc/must-gather-4pq84" podStartSLOduration=4.843583485 podStartE2EDuration="4.843583485s" podCreationTimestamp="2025-10-01 17:05:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 17:05:54.928427041 +0000 UTC m=+4007.829979618" watchObservedRunningTime="2025-10-01 17:05:57.843583485 +0000 UTC m=+4010.745136062" Oct 01 17:05:57 crc kubenswrapper[4726]: I1001 17:05:57.858629 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lb9wc/crc-debug-bdvsm"] Oct 01 17:05:57 crc kubenswrapper[4726]: I1001 17:05:57.859765 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/crc-debug-bdvsm" Oct 01 17:05:57 crc kubenswrapper[4726]: I1001 17:05:57.972287 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e11cc71c-ffab-48e3-a752-ac00955a063f-host\") pod \"crc-debug-bdvsm\" (UID: \"e11cc71c-ffab-48e3-a752-ac00955a063f\") " pod="openshift-must-gather-lb9wc/crc-debug-bdvsm" Oct 01 17:05:57 crc kubenswrapper[4726]: I1001 17:05:57.972381 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49mhb\" (UniqueName: \"kubernetes.io/projected/e11cc71c-ffab-48e3-a752-ac00955a063f-kube-api-access-49mhb\") pod \"crc-debug-bdvsm\" (UID: \"e11cc71c-ffab-48e3-a752-ac00955a063f\") " pod="openshift-must-gather-lb9wc/crc-debug-bdvsm" Oct 01 17:05:58 crc kubenswrapper[4726]: I1001 17:05:58.074064 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e11cc71c-ffab-48e3-a752-ac00955a063f-host\") pod \"crc-debug-bdvsm\" (UID: \"e11cc71c-ffab-48e3-a752-ac00955a063f\") " pod="openshift-must-gather-lb9wc/crc-debug-bdvsm" Oct 01 17:05:58 crc kubenswrapper[4726]: I1001 17:05:58.074124 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49mhb\" (UniqueName: \"kubernetes.io/projected/e11cc71c-ffab-48e3-a752-ac00955a063f-kube-api-access-49mhb\") pod \"crc-debug-bdvsm\" (UID: \"e11cc71c-ffab-48e3-a752-ac00955a063f\") " pod="openshift-must-gather-lb9wc/crc-debug-bdvsm" Oct 01 17:05:58 crc kubenswrapper[4726]: I1001 17:05:58.074262 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e11cc71c-ffab-48e3-a752-ac00955a063f-host\") pod \"crc-debug-bdvsm\" (UID: \"e11cc71c-ffab-48e3-a752-ac00955a063f\") " pod="openshift-must-gather-lb9wc/crc-debug-bdvsm" Oct 01 17:05:58 crc kubenswrapper[4726]: I1001 17:05:58.097361 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49mhb\" (UniqueName: \"kubernetes.io/projected/e11cc71c-ffab-48e3-a752-ac00955a063f-kube-api-access-49mhb\") pod \"crc-debug-bdvsm\" (UID: \"e11cc71c-ffab-48e3-a752-ac00955a063f\") " pod="openshift-must-gather-lb9wc/crc-debug-bdvsm" Oct 01 17:05:58 crc kubenswrapper[4726]: I1001 17:05:58.198991 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/crc-debug-bdvsm" Oct 01 17:05:58 crc kubenswrapper[4726]: W1001 17:05:58.235417 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode11cc71c_ffab_48e3_a752_ac00955a063f.slice/crio-6126cce1ee2a6a7ca59f741025497e992fb7f1fe7ee71c4facef8b85bac1ff93 WatchSource:0}: Error finding container 6126cce1ee2a6a7ca59f741025497e992fb7f1fe7ee71c4facef8b85bac1ff93: Status 404 returned error can't find the container with id 6126cce1ee2a6a7ca59f741025497e992fb7f1fe7ee71c4facef8b85bac1ff93 Oct 01 17:05:58 crc kubenswrapper[4726]: I1001 17:05:58.937184 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lb9wc/crc-debug-bdvsm" event={"ID":"e11cc71c-ffab-48e3-a752-ac00955a063f","Type":"ContainerStarted","Data":"da1d0ba55764b039fb57e7b9556415de4d55e3244253f150cc8aa6c57b476a21"} Oct 01 17:05:58 crc kubenswrapper[4726]: I1001 17:05:58.938665 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lb9wc/crc-debug-bdvsm" event={"ID":"e11cc71c-ffab-48e3-a752-ac00955a063f","Type":"ContainerStarted","Data":"6126cce1ee2a6a7ca59f741025497e992fb7f1fe7ee71c4facef8b85bac1ff93"} Oct 01 17:05:58 crc kubenswrapper[4726]: I1001 17:05:58.958045 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-lb9wc/crc-debug-bdvsm" podStartSLOduration=1.958019505 podStartE2EDuration="1.958019505s" podCreationTimestamp="2025-10-01 17:05:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 17:05:58.951991973 +0000 UTC m=+4011.853544560" watchObservedRunningTime="2025-10-01 17:05:58.958019505 +0000 UTC m=+4011.859572092" Oct 01 17:06:25 crc kubenswrapper[4726]: I1001 17:06:25.466923 4726 scope.go:117] "RemoveContainer" containerID="14c1a933b7e653d9a8218f2bf6a40982eafba5999ff73f7ddb84b1bce7f41d6d" Oct 01 17:07:01 crc kubenswrapper[4726]: I1001 17:07:01.474549 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5649c4d89d-cjbv9_e28be6af-1011-49d4-a949-77fdf75e07d9/barbican-api/0.log" Oct 01 17:07:01 crc kubenswrapper[4726]: I1001 17:07:01.507707 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5649c4d89d-cjbv9_e28be6af-1011-49d4-a949-77fdf75e07d9/barbican-api-log/0.log" Oct 01 17:07:01 crc kubenswrapper[4726]: I1001 17:07:01.644444 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-99545b666-4wmpl_7f25ba81-3f98-4d63-8035-2b30a1df5f54/barbican-keystone-listener/0.log" Oct 01 17:07:01 crc kubenswrapper[4726]: I1001 17:07:01.766226 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-99545b666-4wmpl_7f25ba81-3f98-4d63-8035-2b30a1df5f54/barbican-keystone-listener-log/0.log" Oct 01 17:07:01 crc kubenswrapper[4726]: I1001 17:07:01.871018 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-57cbb4948c-gcc56_94005d17-9bb0-4039-b493-5057a554f054/barbican-worker/0.log" Oct 01 17:07:01 crc kubenswrapper[4726]: I1001 17:07:01.976913 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-57cbb4948c-gcc56_94005d17-9bb0-4039-b493-5057a554f054/barbican-worker-log/0.log" Oct 01 17:07:02 crc kubenswrapper[4726]: I1001 17:07:02.067889 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-vrwzd_bf9f4ecc-09d0-482c-81d2-d7fe9503de1f/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:02 crc kubenswrapper[4726]: I1001 17:07:02.300412 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a303d6a-38ee-4641-b10b-d2627121df03/ceilometer-notification-agent/0.log" Oct 01 17:07:02 crc kubenswrapper[4726]: I1001 17:07:02.318978 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a303d6a-38ee-4641-b10b-d2627121df03/ceilometer-central-agent/0.log" Oct 01 17:07:02 crc kubenswrapper[4726]: I1001 17:07:02.431987 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a303d6a-38ee-4641-b10b-d2627121df03/proxy-httpd/0.log" Oct 01 17:07:02 crc kubenswrapper[4726]: I1001 17:07:02.487419 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a303d6a-38ee-4641-b10b-d2627121df03/sg-core/0.log" Oct 01 17:07:02 crc kubenswrapper[4726]: I1001 17:07:02.667338 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7/cinder-api-log/0.log" Oct 01 17:07:02 crc kubenswrapper[4726]: I1001 17:07:02.739886 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_20d50cfe-7bfc-4295-9c74-7f6dd16eb1c7/cinder-api/0.log" Oct 01 17:07:02 crc kubenswrapper[4726]: I1001 17:07:02.933321 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_aae75086-c497-417c-82aa-9ed0e320b325/cinder-scheduler/0.log" Oct 01 17:07:02 crc kubenswrapper[4726]: I1001 17:07:02.972944 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_aae75086-c497-417c-82aa-9ed0e320b325/probe/0.log" Oct 01 17:07:03 crc kubenswrapper[4726]: I1001 17:07:03.129437 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-fttzz_ec513d51-0278-4d26-aaa7-a66264c32356/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:03 crc kubenswrapper[4726]: I1001 17:07:03.213498 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-4s76w_b8c389e9-bf2c-4e85-a2ec-fc334262347f/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:03 crc kubenswrapper[4726]: I1001 17:07:03.438212 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-tpmnk_98f0cdbd-7ea0-486b-92bc-14e78ab54ffd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:03 crc kubenswrapper[4726]: I1001 17:07:03.518670 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-cx7ps_09379814-0816-4c10-ba61-5d2875e3841a/init/0.log" Oct 01 17:07:03 crc kubenswrapper[4726]: I1001 17:07:03.700689 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-cx7ps_09379814-0816-4c10-ba61-5d2875e3841a/init/0.log" Oct 01 17:07:03 crc kubenswrapper[4726]: I1001 17:07:03.739464 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-cx7ps_09379814-0816-4c10-ba61-5d2875e3841a/dnsmasq-dns/0.log" Oct 01 17:07:03 crc kubenswrapper[4726]: I1001 17:07:03.911936 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-g7c58_bf27af63-5f62-4d0b-98e6-a7a42fb80d7b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:03 crc kubenswrapper[4726]: I1001 17:07:03.991751 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_7aa87400-a85d-4d53-8b65-915138d87c01/glance-httpd/0.log" Oct 01 17:07:04 crc kubenswrapper[4726]: I1001 17:07:04.109676 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_7aa87400-a85d-4d53-8b65-915138d87c01/glance-log/0.log" Oct 01 17:07:04 crc kubenswrapper[4726]: I1001 17:07:04.224285 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b4a301ba-9e78-4262-9efa-3d4d91a5957a/glance-httpd/0.log" Oct 01 17:07:04 crc kubenswrapper[4726]: I1001 17:07:04.365915 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b4a301ba-9e78-4262-9efa-3d4d91a5957a/glance-log/0.log" Oct 01 17:07:04 crc kubenswrapper[4726]: I1001 17:07:04.692931 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7b5b89f8d4-d8p9g_c8ee9afe-8056-4de3-bbff-524420e36290/horizon/0.log" Oct 01 17:07:04 crc kubenswrapper[4726]: I1001 17:07:04.874182 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-tkx4j_d6a1971b-c5b2-4a12-90f2-d986c3ac37b2/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:05 crc kubenswrapper[4726]: I1001 17:07:05.072448 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-qrzm2_4cdcd93d-244a-4adb-b295-ae609306b749/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:05 crc kubenswrapper[4726]: I1001 17:07:05.108170 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7b5b89f8d4-d8p9g_c8ee9afe-8056-4de3-bbff-524420e36290/horizon-log/0.log" Oct 01 17:07:05 crc kubenswrapper[4726]: I1001 17:07:05.280044 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29322301-t47c4_91502936-cb39-48a4-9e82-8cec0e8fb5ab/keystone-cron/0.log" Oct 01 17:07:05 crc kubenswrapper[4726]: I1001 17:07:05.362031 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-75bf4bdcc5-mbcpq_033ea7d7-f7e9-49dd-8c50-5402d4a1aa89/keystone-api/0.log" Oct 01 17:07:05 crc kubenswrapper[4726]: I1001 17:07:05.501291 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_d17a343c-2ff2-4c6e-a224-c8d13f476243/kube-state-metrics/0.log" Oct 01 17:07:05 crc kubenswrapper[4726]: I1001 17:07:05.561404 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-svtfg_a66fccbb-d629-4290-a469-066027643cfa/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:05 crc kubenswrapper[4726]: I1001 17:07:05.889992 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-598fc759b9-28cln_16ed9afc-f760-4a59-8db3-8d76c1bd42fa/neutron-api/0.log" Oct 01 17:07:05 crc kubenswrapper[4726]: I1001 17:07:05.905136 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-598fc759b9-28cln_16ed9afc-f760-4a59-8db3-8d76c1bd42fa/neutron-httpd/0.log" Oct 01 17:07:06 crc kubenswrapper[4726]: I1001 17:07:06.182960 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-4fpd6_ef118af6-8d88-4bdd-bb43-c8532cd50f42/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:06 crc kubenswrapper[4726]: I1001 17:07:06.679411 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6188ef34-0980-4918-bb75-27886226804e/nova-api-log/0.log" Oct 01 17:07:06 crc kubenswrapper[4726]: I1001 17:07:06.875861 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_db204841-b5e7-4ed3-bd88-e6fb7f9d627f/nova-cell0-conductor-conductor/0.log" Oct 01 17:07:07 crc kubenswrapper[4726]: I1001 17:07:07.165255 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6188ef34-0980-4918-bb75-27886226804e/nova-api-api/0.log" Oct 01 17:07:07 crc kubenswrapper[4726]: I1001 17:07:07.700362 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_812fbec4-8c87-4e63-b7e6-e165412dc519/nova-cell1-conductor-conductor/0.log" Oct 01 17:07:07 crc kubenswrapper[4726]: I1001 17:07:07.903010 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_0c2bafde-fa7b-4e17-a5da-e2f2caa63012/nova-cell1-novncproxy-novncproxy/0.log" Oct 01 17:07:08 crc kubenswrapper[4726]: I1001 17:07:08.024583 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-d77vq_d2d4ae0a-aa42-4179-9e7b-cdde9ba99d6d/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:08 crc kubenswrapper[4726]: I1001 17:07:08.190367 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8ec3f7f2-1bb9-47a0-9f62-54625de5bf67/nova-metadata-log/0.log" Oct 01 17:07:08 crc kubenswrapper[4726]: I1001 17:07:08.625979 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c094983e-ab3f-4359-a0ed-5481f349409a/nova-scheduler-scheduler/0.log" Oct 01 17:07:08 crc kubenswrapper[4726]: I1001 17:07:08.808220 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e08387ed-ffb4-4488-9345-02504e1c66cc/mysql-bootstrap/0.log" Oct 01 17:07:09 crc kubenswrapper[4726]: I1001 17:07:09.692640 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e08387ed-ffb4-4488-9345-02504e1c66cc/mysql-bootstrap/0.log" Oct 01 17:07:09 crc kubenswrapper[4726]: I1001 17:07:09.729964 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e08387ed-ffb4-4488-9345-02504e1c66cc/galera/0.log" Oct 01 17:07:09 crc kubenswrapper[4726]: I1001 17:07:09.798608 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8ec3f7f2-1bb9-47a0-9f62-54625de5bf67/nova-metadata-metadata/0.log" Oct 01 17:07:10 crc kubenswrapper[4726]: I1001 17:07:10.001319 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d743678c-e92d-4515-88d0-27e4a5086aac/mysql-bootstrap/0.log" Oct 01 17:07:10 crc kubenswrapper[4726]: I1001 17:07:10.133481 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d743678c-e92d-4515-88d0-27e4a5086aac/mysql-bootstrap/0.log" Oct 01 17:07:10 crc kubenswrapper[4726]: I1001 17:07:10.298077 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d743678c-e92d-4515-88d0-27e4a5086aac/galera/0.log" Oct 01 17:07:10 crc kubenswrapper[4726]: I1001 17:07:10.399071 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_37c371f5-1756-49e1-82ec-40b7cb86f524/openstackclient/0.log" Oct 01 17:07:10 crc kubenswrapper[4726]: I1001 17:07:10.696435 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-h598x_72fa255e-484d-43ab-a837-d31c20e8e4d4/openstack-network-exporter/0.log" Oct 01 17:07:10 crc kubenswrapper[4726]: I1001 17:07:10.839162 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4msxz_525c9172-edd7-4483-b649-128a2d7ac283/ovsdb-server-init/0.log" Oct 01 17:07:11 crc kubenswrapper[4726]: I1001 17:07:11.013491 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4msxz_525c9172-edd7-4483-b649-128a2d7ac283/ovsdb-server-init/0.log" Oct 01 17:07:11 crc kubenswrapper[4726]: I1001 17:07:11.053652 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4msxz_525c9172-edd7-4483-b649-128a2d7ac283/ovs-vswitchd/0.log" Oct 01 17:07:11 crc kubenswrapper[4726]: I1001 17:07:11.109143 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4msxz_525c9172-edd7-4483-b649-128a2d7ac283/ovsdb-server/0.log" Oct 01 17:07:11 crc kubenswrapper[4726]: I1001 17:07:11.273364 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-z4lk4_f67ec6c1-adf7-47f9-855f-fde38e0430c7/ovn-controller/0.log" Oct 01 17:07:11 crc kubenswrapper[4726]: I1001 17:07:11.473258 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-ptlg9_4d73a4ec-57c2-4bab-890c-4d9dfbcc0ec1/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:11 crc kubenswrapper[4726]: I1001 17:07:11.604944 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_cb9896de-b781-451d-8207-093218e40684/openstack-network-exporter/0.log" Oct 01 17:07:11 crc kubenswrapper[4726]: I1001 17:07:11.663628 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_cb9896de-b781-451d-8207-093218e40684/ovn-northd/0.log" Oct 01 17:07:11 crc kubenswrapper[4726]: I1001 17:07:11.788898 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab/openstack-network-exporter/0.log" Oct 01 17:07:11 crc kubenswrapper[4726]: I1001 17:07:11.903513 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9b88b062-70c9-4ea5-9df5-d7eafcc3c6ab/ovsdbserver-nb/0.log" Oct 01 17:07:12 crc kubenswrapper[4726]: I1001 17:07:12.078914 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_23600bcf-22d0-4de1-be32-5adf413748b4/ovsdbserver-sb/0.log" Oct 01 17:07:12 crc kubenswrapper[4726]: I1001 17:07:12.131889 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_23600bcf-22d0-4de1-be32-5adf413748b4/openstack-network-exporter/0.log" Oct 01 17:07:12 crc kubenswrapper[4726]: I1001 17:07:12.327355 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-68f479c67b-gj9j5_f3ada4b2-6da4-4139-ad97-0d45034918aa/placement-api/0.log" Oct 01 17:07:12 crc kubenswrapper[4726]: I1001 17:07:12.521912 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-68f479c67b-gj9j5_f3ada4b2-6da4-4139-ad97-0d45034918aa/placement-log/0.log" Oct 01 17:07:12 crc kubenswrapper[4726]: I1001 17:07:12.565173 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8885e9d5-6b72-496e-b395-d0f94b3b991c/setup-container/0.log" Oct 01 17:07:12 crc kubenswrapper[4726]: I1001 17:07:12.863297 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8885e9d5-6b72-496e-b395-d0f94b3b991c/rabbitmq/0.log" Oct 01 17:07:12 crc kubenswrapper[4726]: I1001 17:07:12.924659 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8885e9d5-6b72-496e-b395-d0f94b3b991c/setup-container/0.log" Oct 01 17:07:13 crc kubenswrapper[4726]: I1001 17:07:13.108144 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b5c68163-e0cf-43e8-8659-260fb0aca643/setup-container/0.log" Oct 01 17:07:13 crc kubenswrapper[4726]: I1001 17:07:13.300609 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b5c68163-e0cf-43e8-8659-260fb0aca643/setup-container/0.log" Oct 01 17:07:13 crc kubenswrapper[4726]: I1001 17:07:13.315404 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b5c68163-e0cf-43e8-8659-260fb0aca643/rabbitmq/0.log" Oct 01 17:07:13 crc kubenswrapper[4726]: I1001 17:07:13.470033 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-kpz4c_016c122e-b1fa-43a1-a806-93533e24b8f6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:13 crc kubenswrapper[4726]: I1001 17:07:13.620527 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-rsl2g_fd3d79a5-76b2-485f-bf48-e4b19fb0ddc3/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:13 crc kubenswrapper[4726]: I1001 17:07:13.773614 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-ff2wn_01436a04-eda0-42cf-a799-29d2bddd197e/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:13 crc kubenswrapper[4726]: I1001 17:07:13.948477 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-6fb8b_1537a5c5-290d-4e30-ba28-f2b90d0d1fe3/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:14 crc kubenswrapper[4726]: I1001 17:07:14.045824 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-mpm5k_e03d54c6-e53e-4837-9d0d-a9183775699e/ssh-known-hosts-edpm-deployment/0.log" Oct 01 17:07:14 crc kubenswrapper[4726]: I1001 17:07:14.299315 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-57c9dff847-vbzrr_85ac310a-31b1-40c9-9724-2c488c0061e2/proxy-server/0.log" Oct 01 17:07:14 crc kubenswrapper[4726]: I1001 17:07:14.350545 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-57c9dff847-vbzrr_85ac310a-31b1-40c9-9724-2c488c0061e2/proxy-httpd/0.log" Oct 01 17:07:14 crc kubenswrapper[4726]: I1001 17:07:14.540813 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-t4hqx_4ab0d371-af59-407e-b596-3c533ad031b8/swift-ring-rebalance/0.log" Oct 01 17:07:14 crc kubenswrapper[4726]: I1001 17:07:14.917412 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/account-auditor/0.log" Oct 01 17:07:14 crc kubenswrapper[4726]: I1001 17:07:14.953217 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/account-reaper/0.log" Oct 01 17:07:15 crc kubenswrapper[4726]: I1001 17:07:15.105624 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/account-replicator/0.log" Oct 01 17:07:15 crc kubenswrapper[4726]: I1001 17:07:15.106874 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/account-server/0.log" Oct 01 17:07:15 crc kubenswrapper[4726]: I1001 17:07:15.181734 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/container-auditor/0.log" Oct 01 17:07:15 crc kubenswrapper[4726]: I1001 17:07:15.298226 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/container-replicator/0.log" Oct 01 17:07:15 crc kubenswrapper[4726]: I1001 17:07:15.382650 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/container-updater/0.log" Oct 01 17:07:15 crc kubenswrapper[4726]: I1001 17:07:15.383152 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/container-server/0.log" Oct 01 17:07:15 crc kubenswrapper[4726]: I1001 17:07:15.535330 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/object-auditor/0.log" Oct 01 17:07:15 crc kubenswrapper[4726]: I1001 17:07:15.590661 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/object-replicator/0.log" Oct 01 17:07:15 crc kubenswrapper[4726]: I1001 17:07:15.619693 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/object-expirer/0.log" Oct 01 17:07:15 crc kubenswrapper[4726]: I1001 17:07:15.698640 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/object-server/0.log" Oct 01 17:07:15 crc kubenswrapper[4726]: I1001 17:07:15.798590 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/object-updater/0.log" Oct 01 17:07:15 crc kubenswrapper[4726]: I1001 17:07:15.860467 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/rsync/0.log" Oct 01 17:07:15 crc kubenswrapper[4726]: I1001 17:07:15.903090 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6d192df1-7269-4493-b17b-4037addc226e/swift-recon-cron/0.log" Oct 01 17:07:16 crc kubenswrapper[4726]: I1001 17:07:16.145940 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-qwq7k_376afdd4-d312-464e-9627-42656734a785/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:16 crc kubenswrapper[4726]: I1001 17:07:16.280810 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_261dc564-2126-415f-a5ce-a988af8a053c/tempest-tests-tempest-tests-runner/0.log" Oct 01 17:07:16 crc kubenswrapper[4726]: I1001 17:07:16.352656 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_c178394b-ef0f-48ec-992f-58cb68ef7ca6/test-operator-logs-container/0.log" Oct 01 17:07:16 crc kubenswrapper[4726]: I1001 17:07:16.985431 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-l5hwk_c69a5dd3-367d-4db8-907c-cb7afb6da6a9/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 17:07:23 crc kubenswrapper[4726]: I1001 17:07:23.413005 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 17:07:23 crc kubenswrapper[4726]: I1001 17:07:23.413640 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 17:07:24 crc kubenswrapper[4726]: I1001 17:07:24.752931 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fd24681e-73eb-49d0-bd62-4d11a485c18e/memcached/0.log" Oct 01 17:07:52 crc kubenswrapper[4726]: I1001 17:07:52.009001 4726 generic.go:334] "Generic (PLEG): container finished" podID="e11cc71c-ffab-48e3-a752-ac00955a063f" containerID="da1d0ba55764b039fb57e7b9556415de4d55e3244253f150cc8aa6c57b476a21" exitCode=0 Oct 01 17:07:52 crc kubenswrapper[4726]: I1001 17:07:52.009088 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lb9wc/crc-debug-bdvsm" event={"ID":"e11cc71c-ffab-48e3-a752-ac00955a063f","Type":"ContainerDied","Data":"da1d0ba55764b039fb57e7b9556415de4d55e3244253f150cc8aa6c57b476a21"} Oct 01 17:07:53 crc kubenswrapper[4726]: I1001 17:07:53.173439 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/crc-debug-bdvsm" Oct 01 17:07:53 crc kubenswrapper[4726]: I1001 17:07:53.216208 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lb9wc/crc-debug-bdvsm"] Oct 01 17:07:53 crc kubenswrapper[4726]: I1001 17:07:53.224451 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lb9wc/crc-debug-bdvsm"] Oct 01 17:07:53 crc kubenswrapper[4726]: I1001 17:07:53.236588 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49mhb\" (UniqueName: \"kubernetes.io/projected/e11cc71c-ffab-48e3-a752-ac00955a063f-kube-api-access-49mhb\") pod \"e11cc71c-ffab-48e3-a752-ac00955a063f\" (UID: \"e11cc71c-ffab-48e3-a752-ac00955a063f\") " Oct 01 17:07:53 crc kubenswrapper[4726]: I1001 17:07:53.236651 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e11cc71c-ffab-48e3-a752-ac00955a063f-host\") pod \"e11cc71c-ffab-48e3-a752-ac00955a063f\" (UID: \"e11cc71c-ffab-48e3-a752-ac00955a063f\") " Oct 01 17:07:53 crc kubenswrapper[4726]: I1001 17:07:53.236879 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e11cc71c-ffab-48e3-a752-ac00955a063f-host" (OuterVolumeSpecName: "host") pod "e11cc71c-ffab-48e3-a752-ac00955a063f" (UID: "e11cc71c-ffab-48e3-a752-ac00955a063f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 17:07:53 crc kubenswrapper[4726]: I1001 17:07:53.237520 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e11cc71c-ffab-48e3-a752-ac00955a063f-host\") on node \"crc\" DevicePath \"\"" Oct 01 17:07:53 crc kubenswrapper[4726]: I1001 17:07:53.242910 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e11cc71c-ffab-48e3-a752-ac00955a063f-kube-api-access-49mhb" (OuterVolumeSpecName: "kube-api-access-49mhb") pod "e11cc71c-ffab-48e3-a752-ac00955a063f" (UID: "e11cc71c-ffab-48e3-a752-ac00955a063f"). InnerVolumeSpecName "kube-api-access-49mhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 17:07:53 crc kubenswrapper[4726]: I1001 17:07:53.339213 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49mhb\" (UniqueName: \"kubernetes.io/projected/e11cc71c-ffab-48e3-a752-ac00955a063f-kube-api-access-49mhb\") on node \"crc\" DevicePath \"\"" Oct 01 17:07:53 crc kubenswrapper[4726]: I1001 17:07:53.414033 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 17:07:53 crc kubenswrapper[4726]: I1001 17:07:53.414206 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 17:07:53 crc kubenswrapper[4726]: I1001 17:07:53.823351 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e11cc71c-ffab-48e3-a752-ac00955a063f" path="/var/lib/kubelet/pods/e11cc71c-ffab-48e3-a752-ac00955a063f/volumes" Oct 01 17:07:54 crc kubenswrapper[4726]: I1001 17:07:54.029894 4726 scope.go:117] "RemoveContainer" containerID="da1d0ba55764b039fb57e7b9556415de4d55e3244253f150cc8aa6c57b476a21" Oct 01 17:07:54 crc kubenswrapper[4726]: I1001 17:07:54.030053 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/crc-debug-bdvsm" Oct 01 17:07:54 crc kubenswrapper[4726]: I1001 17:07:54.381444 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lb9wc/crc-debug-wt7gg"] Oct 01 17:07:54 crc kubenswrapper[4726]: E1001 17:07:54.382158 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e11cc71c-ffab-48e3-a752-ac00955a063f" containerName="container-00" Oct 01 17:07:54 crc kubenswrapper[4726]: I1001 17:07:54.382172 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e11cc71c-ffab-48e3-a752-ac00955a063f" containerName="container-00" Oct 01 17:07:54 crc kubenswrapper[4726]: I1001 17:07:54.382541 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e11cc71c-ffab-48e3-a752-ac00955a063f" containerName="container-00" Oct 01 17:07:54 crc kubenswrapper[4726]: I1001 17:07:54.383414 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" Oct 01 17:07:54 crc kubenswrapper[4726]: I1001 17:07:54.457571 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc1d797c-ec4e-465b-bbb3-5abaf655548f-host\") pod \"crc-debug-wt7gg\" (UID: \"fc1d797c-ec4e-465b-bbb3-5abaf655548f\") " pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" Oct 01 17:07:54 crc kubenswrapper[4726]: I1001 17:07:54.457801 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64858\" (UniqueName: \"kubernetes.io/projected/fc1d797c-ec4e-465b-bbb3-5abaf655548f-kube-api-access-64858\") pod \"crc-debug-wt7gg\" (UID: \"fc1d797c-ec4e-465b-bbb3-5abaf655548f\") " pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" Oct 01 17:07:54 crc kubenswrapper[4726]: I1001 17:07:54.560251 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64858\" (UniqueName: \"kubernetes.io/projected/fc1d797c-ec4e-465b-bbb3-5abaf655548f-kube-api-access-64858\") pod \"crc-debug-wt7gg\" (UID: \"fc1d797c-ec4e-465b-bbb3-5abaf655548f\") " pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" Oct 01 17:07:54 crc kubenswrapper[4726]: I1001 17:07:54.560616 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc1d797c-ec4e-465b-bbb3-5abaf655548f-host\") pod \"crc-debug-wt7gg\" (UID: \"fc1d797c-ec4e-465b-bbb3-5abaf655548f\") " pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" Oct 01 17:07:54 crc kubenswrapper[4726]: I1001 17:07:54.560729 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc1d797c-ec4e-465b-bbb3-5abaf655548f-host\") pod \"crc-debug-wt7gg\" (UID: \"fc1d797c-ec4e-465b-bbb3-5abaf655548f\") " pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" Oct 01 17:07:54 crc kubenswrapper[4726]: I1001 17:07:54.587336 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64858\" (UniqueName: \"kubernetes.io/projected/fc1d797c-ec4e-465b-bbb3-5abaf655548f-kube-api-access-64858\") pod \"crc-debug-wt7gg\" (UID: \"fc1d797c-ec4e-465b-bbb3-5abaf655548f\") " pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" Oct 01 17:07:54 crc kubenswrapper[4726]: I1001 17:07:54.714302 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" Oct 01 17:07:55 crc kubenswrapper[4726]: I1001 17:07:55.051736 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" event={"ID":"fc1d797c-ec4e-465b-bbb3-5abaf655548f","Type":"ContainerStarted","Data":"d03d8dc429f9663e9d31eb31c8c06d51f1d30d9adbdcfcdb99f09403c10db5fe"} Oct 01 17:07:55 crc kubenswrapper[4726]: I1001 17:07:55.052018 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" event={"ID":"fc1d797c-ec4e-465b-bbb3-5abaf655548f","Type":"ContainerStarted","Data":"974e2cb07b4dda0fcdd68979c431e230b6507c09c98eab53c129212468715f93"} Oct 01 17:07:55 crc kubenswrapper[4726]: I1001 17:07:55.078091 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" podStartSLOduration=1.078056723 podStartE2EDuration="1.078056723s" podCreationTimestamp="2025-10-01 17:07:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 17:07:55.066640136 +0000 UTC m=+4127.968192713" watchObservedRunningTime="2025-10-01 17:07:55.078056723 +0000 UTC m=+4127.979609300" Oct 01 17:07:56 crc kubenswrapper[4726]: I1001 17:07:56.063360 4726 generic.go:334] "Generic (PLEG): container finished" podID="fc1d797c-ec4e-465b-bbb3-5abaf655548f" containerID="d03d8dc429f9663e9d31eb31c8c06d51f1d30d9adbdcfcdb99f09403c10db5fe" exitCode=0 Oct 01 17:07:56 crc kubenswrapper[4726]: I1001 17:07:56.063401 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" event={"ID":"fc1d797c-ec4e-465b-bbb3-5abaf655548f","Type":"ContainerDied","Data":"d03d8dc429f9663e9d31eb31c8c06d51f1d30d9adbdcfcdb99f09403c10db5fe"} Oct 01 17:07:57 crc kubenswrapper[4726]: I1001 17:07:57.174095 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" Oct 01 17:07:57 crc kubenswrapper[4726]: I1001 17:07:57.298768 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64858\" (UniqueName: \"kubernetes.io/projected/fc1d797c-ec4e-465b-bbb3-5abaf655548f-kube-api-access-64858\") pod \"fc1d797c-ec4e-465b-bbb3-5abaf655548f\" (UID: \"fc1d797c-ec4e-465b-bbb3-5abaf655548f\") " Oct 01 17:07:57 crc kubenswrapper[4726]: I1001 17:07:57.298902 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc1d797c-ec4e-465b-bbb3-5abaf655548f-host\") pod \"fc1d797c-ec4e-465b-bbb3-5abaf655548f\" (UID: \"fc1d797c-ec4e-465b-bbb3-5abaf655548f\") " Oct 01 17:07:57 crc kubenswrapper[4726]: I1001 17:07:57.299131 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc1d797c-ec4e-465b-bbb3-5abaf655548f-host" (OuterVolumeSpecName: "host") pod "fc1d797c-ec4e-465b-bbb3-5abaf655548f" (UID: "fc1d797c-ec4e-465b-bbb3-5abaf655548f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 17:07:57 crc kubenswrapper[4726]: I1001 17:07:57.299468 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc1d797c-ec4e-465b-bbb3-5abaf655548f-host\") on node \"crc\" DevicePath \"\"" Oct 01 17:07:57 crc kubenswrapper[4726]: I1001 17:07:57.304495 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc1d797c-ec4e-465b-bbb3-5abaf655548f-kube-api-access-64858" (OuterVolumeSpecName: "kube-api-access-64858") pod "fc1d797c-ec4e-465b-bbb3-5abaf655548f" (UID: "fc1d797c-ec4e-465b-bbb3-5abaf655548f"). InnerVolumeSpecName "kube-api-access-64858". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 17:07:57 crc kubenswrapper[4726]: I1001 17:07:57.400774 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64858\" (UniqueName: \"kubernetes.io/projected/fc1d797c-ec4e-465b-bbb3-5abaf655548f-kube-api-access-64858\") on node \"crc\" DevicePath \"\"" Oct 01 17:07:58 crc kubenswrapper[4726]: I1001 17:07:58.080461 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" event={"ID":"fc1d797c-ec4e-465b-bbb3-5abaf655548f","Type":"ContainerDied","Data":"974e2cb07b4dda0fcdd68979c431e230b6507c09c98eab53c129212468715f93"} Oct 01 17:07:58 crc kubenswrapper[4726]: I1001 17:07:58.080866 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="974e2cb07b4dda0fcdd68979c431e230b6507c09c98eab53c129212468715f93" Oct 01 17:07:58 crc kubenswrapper[4726]: I1001 17:07:58.081282 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/crc-debug-wt7gg" Oct 01 17:08:02 crc kubenswrapper[4726]: I1001 17:08:02.439126 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lb9wc/crc-debug-wt7gg"] Oct 01 17:08:02 crc kubenswrapper[4726]: I1001 17:08:02.448439 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lb9wc/crc-debug-wt7gg"] Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.661353 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lb9wc/crc-debug-nqsmd"] Oct 01 17:08:03 crc kubenswrapper[4726]: E1001 17:08:03.661707 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc1d797c-ec4e-465b-bbb3-5abaf655548f" containerName="container-00" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.661719 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc1d797c-ec4e-465b-bbb3-5abaf655548f" containerName="container-00" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.661969 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc1d797c-ec4e-465b-bbb3-5abaf655548f" containerName="container-00" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.662620 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/crc-debug-nqsmd" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.703082 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-49xhx"] Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.706209 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.714246 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-49xhx"] Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.804793 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-catalog-content\") pod \"certified-operators-49xhx\" (UID: \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\") " pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.804843 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-utilities\") pod \"certified-operators-49xhx\" (UID: \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\") " pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.804872 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edca44cc-3040-4598-9e7c-ed7a7b235771-host\") pod \"crc-debug-nqsmd\" (UID: \"edca44cc-3040-4598-9e7c-ed7a7b235771\") " pod="openshift-must-gather-lb9wc/crc-debug-nqsmd" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.804898 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phlsj\" (UniqueName: \"kubernetes.io/projected/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-kube-api-access-phlsj\") pod \"certified-operators-49xhx\" (UID: \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\") " pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.804999 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwrbn\" (UniqueName: \"kubernetes.io/projected/edca44cc-3040-4598-9e7c-ed7a7b235771-kube-api-access-hwrbn\") pod \"crc-debug-nqsmd\" (UID: \"edca44cc-3040-4598-9e7c-ed7a7b235771\") " pod="openshift-must-gather-lb9wc/crc-debug-nqsmd" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.819281 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc1d797c-ec4e-465b-bbb3-5abaf655548f" path="/var/lib/kubelet/pods/fc1d797c-ec4e-465b-bbb3-5abaf655548f/volumes" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.906625 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phlsj\" (UniqueName: \"kubernetes.io/projected/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-kube-api-access-phlsj\") pod \"certified-operators-49xhx\" (UID: \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\") " pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.906833 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwrbn\" (UniqueName: \"kubernetes.io/projected/edca44cc-3040-4598-9e7c-ed7a7b235771-kube-api-access-hwrbn\") pod \"crc-debug-nqsmd\" (UID: \"edca44cc-3040-4598-9e7c-ed7a7b235771\") " pod="openshift-must-gather-lb9wc/crc-debug-nqsmd" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.906887 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-catalog-content\") pod \"certified-operators-49xhx\" (UID: \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\") " pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.906924 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-utilities\") pod \"certified-operators-49xhx\" (UID: \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\") " pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.906959 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edca44cc-3040-4598-9e7c-ed7a7b235771-host\") pod \"crc-debug-nqsmd\" (UID: \"edca44cc-3040-4598-9e7c-ed7a7b235771\") " pod="openshift-must-gather-lb9wc/crc-debug-nqsmd" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.907094 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edca44cc-3040-4598-9e7c-ed7a7b235771-host\") pod \"crc-debug-nqsmd\" (UID: \"edca44cc-3040-4598-9e7c-ed7a7b235771\") " pod="openshift-must-gather-lb9wc/crc-debug-nqsmd" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.907495 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-catalog-content\") pod \"certified-operators-49xhx\" (UID: \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\") " pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.907519 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-utilities\") pod \"certified-operators-49xhx\" (UID: \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\") " pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.929766 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwrbn\" (UniqueName: \"kubernetes.io/projected/edca44cc-3040-4598-9e7c-ed7a7b235771-kube-api-access-hwrbn\") pod \"crc-debug-nqsmd\" (UID: \"edca44cc-3040-4598-9e7c-ed7a7b235771\") " pod="openshift-must-gather-lb9wc/crc-debug-nqsmd" Oct 01 17:08:03 crc kubenswrapper[4726]: I1001 17:08:03.929875 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phlsj\" (UniqueName: \"kubernetes.io/projected/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-kube-api-access-phlsj\") pod \"certified-operators-49xhx\" (UID: \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\") " pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:04 crc kubenswrapper[4726]: I1001 17:08:04.002421 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/crc-debug-nqsmd" Oct 01 17:08:04 crc kubenswrapper[4726]: I1001 17:08:04.034165 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:04 crc kubenswrapper[4726]: I1001 17:08:04.158420 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lb9wc/crc-debug-nqsmd" event={"ID":"edca44cc-3040-4598-9e7c-ed7a7b235771","Type":"ContainerStarted","Data":"8bbbacb3289e8206fc301dea07133d6b126e40f6446856768445c17399480787"} Oct 01 17:08:04 crc kubenswrapper[4726]: I1001 17:08:04.784926 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-49xhx"] Oct 01 17:08:04 crc kubenswrapper[4726]: W1001 17:08:04.791452 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f46f2e2_93c0_49a3_b901_3bd9e06f6342.slice/crio-835fe2fc2f70d6e6243f0aa0f68c380ba29fc3fb1412ef2c434fb2e27c1054a4 WatchSource:0}: Error finding container 835fe2fc2f70d6e6243f0aa0f68c380ba29fc3fb1412ef2c434fb2e27c1054a4: Status 404 returned error can't find the container with id 835fe2fc2f70d6e6243f0aa0f68c380ba29fc3fb1412ef2c434fb2e27c1054a4 Oct 01 17:08:05 crc kubenswrapper[4726]: I1001 17:08:05.167520 4726 generic.go:334] "Generic (PLEG): container finished" podID="edca44cc-3040-4598-9e7c-ed7a7b235771" containerID="080aaaefeb2efb3918518bdf8175f75c951e829c660f6d7b39b67c7e6263ad31" exitCode=0 Oct 01 17:08:05 crc kubenswrapper[4726]: I1001 17:08:05.167589 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lb9wc/crc-debug-nqsmd" event={"ID":"edca44cc-3040-4598-9e7c-ed7a7b235771","Type":"ContainerDied","Data":"080aaaefeb2efb3918518bdf8175f75c951e829c660f6d7b39b67c7e6263ad31"} Oct 01 17:08:05 crc kubenswrapper[4726]: I1001 17:08:05.169587 4726 generic.go:334] "Generic (PLEG): container finished" podID="4f46f2e2-93c0-49a3-b901-3bd9e06f6342" containerID="41966207ce8b2883fbdff899448e9ca3c40457e02017ac2d4bac3081022c4117" exitCode=0 Oct 01 17:08:05 crc kubenswrapper[4726]: I1001 17:08:05.169692 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49xhx" event={"ID":"4f46f2e2-93c0-49a3-b901-3bd9e06f6342","Type":"ContainerDied","Data":"41966207ce8b2883fbdff899448e9ca3c40457e02017ac2d4bac3081022c4117"} Oct 01 17:08:05 crc kubenswrapper[4726]: I1001 17:08:05.169770 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49xhx" event={"ID":"4f46f2e2-93c0-49a3-b901-3bd9e06f6342","Type":"ContainerStarted","Data":"835fe2fc2f70d6e6243f0aa0f68c380ba29fc3fb1412ef2c434fb2e27c1054a4"} Oct 01 17:08:05 crc kubenswrapper[4726]: I1001 17:08:05.223391 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lb9wc/crc-debug-nqsmd"] Oct 01 17:08:05 crc kubenswrapper[4726]: I1001 17:08:05.230822 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lb9wc/crc-debug-nqsmd"] Oct 01 17:08:06 crc kubenswrapper[4726]: I1001 17:08:06.887838 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/crc-debug-nqsmd" Oct 01 17:08:06 crc kubenswrapper[4726]: I1001 17:08:06.978173 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwrbn\" (UniqueName: \"kubernetes.io/projected/edca44cc-3040-4598-9e7c-ed7a7b235771-kube-api-access-hwrbn\") pod \"edca44cc-3040-4598-9e7c-ed7a7b235771\" (UID: \"edca44cc-3040-4598-9e7c-ed7a7b235771\") " Oct 01 17:08:06 crc kubenswrapper[4726]: I1001 17:08:06.978542 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edca44cc-3040-4598-9e7c-ed7a7b235771-host\") pod \"edca44cc-3040-4598-9e7c-ed7a7b235771\" (UID: \"edca44cc-3040-4598-9e7c-ed7a7b235771\") " Oct 01 17:08:06 crc kubenswrapper[4726]: I1001 17:08:06.979163 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/edca44cc-3040-4598-9e7c-ed7a7b235771-host" (OuterVolumeSpecName: "host") pod "edca44cc-3040-4598-9e7c-ed7a7b235771" (UID: "edca44cc-3040-4598-9e7c-ed7a7b235771"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 17:08:06 crc kubenswrapper[4726]: I1001 17:08:06.979566 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edca44cc-3040-4598-9e7c-ed7a7b235771-host\") on node \"crc\" DevicePath \"\"" Oct 01 17:08:06 crc kubenswrapper[4726]: I1001 17:08:06.986350 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edca44cc-3040-4598-9e7c-ed7a7b235771-kube-api-access-hwrbn" (OuterVolumeSpecName: "kube-api-access-hwrbn") pod "edca44cc-3040-4598-9e7c-ed7a7b235771" (UID: "edca44cc-3040-4598-9e7c-ed7a7b235771"). InnerVolumeSpecName "kube-api-access-hwrbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 17:08:07 crc kubenswrapper[4726]: I1001 17:08:07.080547 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwrbn\" (UniqueName: \"kubernetes.io/projected/edca44cc-3040-4598-9e7c-ed7a7b235771-kube-api-access-hwrbn\") on node \"crc\" DevicePath \"\"" Oct 01 17:08:07 crc kubenswrapper[4726]: I1001 17:08:07.194305 4726 scope.go:117] "RemoveContainer" containerID="080aaaefeb2efb3918518bdf8175f75c951e829c660f6d7b39b67c7e6263ad31" Oct 01 17:08:07 crc kubenswrapper[4726]: I1001 17:08:07.194343 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/crc-debug-nqsmd" Oct 01 17:08:07 crc kubenswrapper[4726]: I1001 17:08:07.197581 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49xhx" event={"ID":"4f46f2e2-93c0-49a3-b901-3bd9e06f6342","Type":"ContainerStarted","Data":"f39287131613603a4273ea8139fe23ec6851aa1884b5023cf2545c48774fd571"} Oct 01 17:08:07 crc kubenswrapper[4726]: I1001 17:08:07.515854 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd_6eef09a6-cce6-45d8-ac7e-6b09f74d730a/util/0.log" Oct 01 17:08:07 crc kubenswrapper[4726]: I1001 17:08:07.734804 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd_6eef09a6-cce6-45d8-ac7e-6b09f74d730a/util/0.log" Oct 01 17:08:07 crc kubenswrapper[4726]: I1001 17:08:07.777859 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd_6eef09a6-cce6-45d8-ac7e-6b09f74d730a/pull/0.log" Oct 01 17:08:07 crc kubenswrapper[4726]: I1001 17:08:07.783862 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd_6eef09a6-cce6-45d8-ac7e-6b09f74d730a/pull/0.log" Oct 01 17:08:07 crc kubenswrapper[4726]: I1001 17:08:07.844358 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edca44cc-3040-4598-9e7c-ed7a7b235771" path="/var/lib/kubelet/pods/edca44cc-3040-4598-9e7c-ed7a7b235771/volumes" Oct 01 17:08:07 crc kubenswrapper[4726]: I1001 17:08:07.954207 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd_6eef09a6-cce6-45d8-ac7e-6b09f74d730a/util/0.log" Oct 01 17:08:08 crc kubenswrapper[4726]: I1001 17:08:08.013502 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd_6eef09a6-cce6-45d8-ac7e-6b09f74d730a/extract/0.log" Oct 01 17:08:08 crc kubenswrapper[4726]: I1001 17:08:08.014085 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62f343461bb1f516c3387e2699981d370ec71b3ff7bab8ccc23c973b734p7nd_6eef09a6-cce6-45d8-ac7e-6b09f74d730a/pull/0.log" Oct 01 17:08:08 crc kubenswrapper[4726]: I1001 17:08:08.110262 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-wckhv_85f77f5d-6b74-40b5-b238-d3515b4a685c/kube-rbac-proxy/0.log" Oct 01 17:08:08 crc kubenswrapper[4726]: I1001 17:08:08.922183 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-zxd4h_117d55f2-eee8-454b-9091-3ffffd61c547/kube-rbac-proxy/0.log" Oct 01 17:08:08 crc kubenswrapper[4726]: I1001 17:08:08.968471 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-wckhv_85f77f5d-6b74-40b5-b238-d3515b4a685c/manager/0.log" Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.011498 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-zxd4h_117d55f2-eee8-454b-9091-3ffffd61c547/manager/0.log" Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.096514 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-wgkz2_71d9bfcf-89ac-440f-b2b1-1c2969859b0d/kube-rbac-proxy/0.log" Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.139431 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-wgkz2_71d9bfcf-89ac-440f-b2b1-1c2969859b0d/manager/0.log" Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.222897 4726 generic.go:334] "Generic (PLEG): container finished" podID="4f46f2e2-93c0-49a3-b901-3bd9e06f6342" containerID="f39287131613603a4273ea8139fe23ec6851aa1884b5023cf2545c48774fd571" exitCode=0 Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.222939 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49xhx" event={"ID":"4f46f2e2-93c0-49a3-b901-3bd9e06f6342","Type":"ContainerDied","Data":"f39287131613603a4273ea8139fe23ec6851aa1884b5023cf2545c48774fd571"} Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.342931 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-67rj9_daf7d005-ec55-4cdb-91d6-eac465e42e17/kube-rbac-proxy/0.log" Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.397505 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-qlb92_f82a5172-6387-4a24-a7e4-a5bd4c90d310/kube-rbac-proxy/0.log" Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.419675 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-67rj9_daf7d005-ec55-4cdb-91d6-eac465e42e17/manager/0.log" Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.511515 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-qlb92_f82a5172-6387-4a24-a7e4-a5bd4c90d310/manager/0.log" Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.601755 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-zjwvl_39860f7b-9a7f-44b1-8bf4-7a2204092c18/kube-rbac-proxy/0.log" Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.604294 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-zjwvl_39860f7b-9a7f-44b1-8bf4-7a2204092c18/manager/0.log" Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.733594 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-z5lfd_6b63c738-b108-4193-9c9d-11f4eb8227aa/kube-rbac-proxy/0.log" Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.911900 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-z5lfd_6b63c738-b108-4193-9c9d-11f4eb8227aa/manager/0.log" Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.939996 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-cpbbq_8e085c54-533c-4186-a6fe-5d8a5ccbac2f/manager/0.log" Oct 01 17:08:09 crc kubenswrapper[4726]: I1001 17:08:09.942162 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-cpbbq_8e085c54-533c-4186-a6fe-5d8a5ccbac2f/kube-rbac-proxy/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.056416 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-5hvl7_23922646-bd87-424b-85d7-8a2a55056438/kube-rbac-proxy/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.165621 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-5hvl7_23922646-bd87-424b-85d7-8a2a55056438/manager/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.172390 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-qd8pc_2d90550d-aedc-4e50-81f9-da6f285f8c2b/kube-rbac-proxy/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.231878 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49xhx" event={"ID":"4f46f2e2-93c0-49a3-b901-3bd9e06f6342","Type":"ContainerStarted","Data":"08bf037d93bf1c62583bb87d7f3c9616d25e86d83a86e4f213901670480b6a7f"} Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.255888 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-49xhx" podStartSLOduration=2.764230575 podStartE2EDuration="7.255870639s" podCreationTimestamp="2025-10-01 17:08:03 +0000 UTC" firstStartedPulling="2025-10-01 17:08:05.171710985 +0000 UTC m=+4138.073263602" lastFinishedPulling="2025-10-01 17:08:09.663351089 +0000 UTC m=+4142.564903666" observedRunningTime="2025-10-01 17:08:10.247709935 +0000 UTC m=+4143.149262512" watchObservedRunningTime="2025-10-01 17:08:10.255870639 +0000 UTC m=+4143.157423206" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.260885 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-qd8pc_2d90550d-aedc-4e50-81f9-da6f285f8c2b/manager/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.388249 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-5vlmb_9f6777bb-5041-4c66-967a-069fa217bf99/kube-rbac-proxy/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.399957 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-5vlmb_9f6777bb-5041-4c66-967a-069fa217bf99/manager/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.484341 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-w9xkn_0d89451f-e13a-4195-a2b2-f16443c43463/kube-rbac-proxy/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.579077 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-w9xkn_0d89451f-e13a-4195-a2b2-f16443c43463/manager/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.596998 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-zdkvq_0c8fbaaf-4ce0-4388-855e-fb85534f05e6/kube-rbac-proxy/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.662168 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-zdkvq_0c8fbaaf-4ce0-4388-855e-fb85534f05e6/manager/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.780088 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-t2vwz_b9f1f46c-c6fa-443f-b53d-54614779c0cc/kube-rbac-proxy/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.808019 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-t2vwz_b9f1f46c-c6fa-443f-b53d-54614779c0cc/manager/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.868882 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf_420c55b8-4442-41fa-98a4-338f7338ceb4/kube-rbac-proxy/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.901656 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-77b9676b8cfgqhf_420c55b8-4442-41fa-98a4-338f7338ceb4/manager/0.log" Oct 01 17:08:10 crc kubenswrapper[4726]: I1001 17:08:10.992177 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5f569c4895-tl4ld_42f6c148-bcc5-4054-88da-dc79d49baeff/kube-rbac-proxy/0.log" Oct 01 17:08:11 crc kubenswrapper[4726]: I1001 17:08:11.159560 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6877df559-w8rr9_07c756cd-913e-4237-a8eb-06dde60b6d4b/kube-rbac-proxy/0.log" Oct 01 17:08:11 crc kubenswrapper[4726]: I1001 17:08:11.329561 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6877df559-w8rr9_07c756cd-913e-4237-a8eb-06dde60b6d4b/operator/0.log" Oct 01 17:08:11 crc kubenswrapper[4726]: I1001 17:08:11.374232 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-wh8qd_62b6b99f-a44a-445b-9183-917d39934259/registry-server/0.log" Oct 01 17:08:11 crc kubenswrapper[4726]: I1001 17:08:11.587873 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-j4t5g_1845e7cd-8784-4994-8f4f-3015bf7d6b3b/kube-rbac-proxy/0.log" Oct 01 17:08:11 crc kubenswrapper[4726]: I1001 17:08:11.612862 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-j4t5g_1845e7cd-8784-4994-8f4f-3015bf7d6b3b/manager/0.log" Oct 01 17:08:11 crc kubenswrapper[4726]: I1001 17:08:11.658987 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-7s6mh_ff753ea2-fa5b-4367-b476-596fd49d0557/kube-rbac-proxy/0.log" Oct 01 17:08:11 crc kubenswrapper[4726]: I1001 17:08:11.832983 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-7s6mh_ff753ea2-fa5b-4367-b476-596fd49d0557/manager/0.log" Oct 01 17:08:11 crc kubenswrapper[4726]: I1001 17:08:11.860121 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-f4rbv_965ea363-f11c-4ced-9988-debe220ee3f9/operator/0.log" Oct 01 17:08:12 crc kubenswrapper[4726]: I1001 17:08:12.087469 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-pl7jv_fa99155f-85e2-406e-84c4-4a95ab74407f/manager/0.log" Oct 01 17:08:12 crc kubenswrapper[4726]: I1001 17:08:12.119002 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5f569c4895-tl4ld_42f6c148-bcc5-4054-88da-dc79d49baeff/manager/0.log" Oct 01 17:08:12 crc kubenswrapper[4726]: I1001 17:08:12.128718 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-pl7jv_fa99155f-85e2-406e-84c4-4a95ab74407f/kube-rbac-proxy/0.log" Oct 01 17:08:12 crc kubenswrapper[4726]: I1001 17:08:12.138454 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7866c8b47c-z6rlg_75c9f506-beb2-4fe6-9ada-3ed684131102/kube-rbac-proxy/0.log" Oct 01 17:08:12 crc kubenswrapper[4726]: I1001 17:08:12.311980 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-nv4ww_0396f184-2150-4944-a81c-c0d0fafe1317/kube-rbac-proxy/0.log" Oct 01 17:08:12 crc kubenswrapper[4726]: I1001 17:08:12.344042 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-nv4ww_0396f184-2150-4944-a81c-c0d0fafe1317/manager/0.log" Oct 01 17:08:12 crc kubenswrapper[4726]: I1001 17:08:12.358035 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7866c8b47c-z6rlg_75c9f506-beb2-4fe6-9ada-3ed684131102/manager/0.log" Oct 01 17:08:12 crc kubenswrapper[4726]: I1001 17:08:12.468877 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-9nqm2_b824be4d-8e37-42b5-8fe2-0dc6a1c4b4a1/kube-rbac-proxy/0.log" Oct 01 17:08:12 crc kubenswrapper[4726]: I1001 17:08:12.514401 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-9nqm2_b824be4d-8e37-42b5-8fe2-0dc6a1c4b4a1/manager/0.log" Oct 01 17:08:14 crc kubenswrapper[4726]: I1001 17:08:14.034820 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:14 crc kubenswrapper[4726]: I1001 17:08:14.035189 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:14 crc kubenswrapper[4726]: I1001 17:08:14.091769 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:14 crc kubenswrapper[4726]: I1001 17:08:14.325721 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:14 crc kubenswrapper[4726]: I1001 17:08:14.387499 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-49xhx"] Oct 01 17:08:16 crc kubenswrapper[4726]: I1001 17:08:16.290076 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-49xhx" podUID="4f46f2e2-93c0-49a3-b901-3bd9e06f6342" containerName="registry-server" containerID="cri-o://08bf037d93bf1c62583bb87d7f3c9616d25e86d83a86e4f213901670480b6a7f" gracePeriod=2 Oct 01 17:08:16 crc kubenswrapper[4726]: I1001 17:08:16.736234 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:16 crc kubenswrapper[4726]: I1001 17:08:16.862979 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-catalog-content\") pod \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\" (UID: \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\") " Oct 01 17:08:16 crc kubenswrapper[4726]: I1001 17:08:16.863123 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phlsj\" (UniqueName: \"kubernetes.io/projected/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-kube-api-access-phlsj\") pod \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\" (UID: \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\") " Oct 01 17:08:16 crc kubenswrapper[4726]: I1001 17:08:16.863312 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-utilities\") pod \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\" (UID: \"4f46f2e2-93c0-49a3-b901-3bd9e06f6342\") " Oct 01 17:08:16 crc kubenswrapper[4726]: I1001 17:08:16.864678 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-utilities" (OuterVolumeSpecName: "utilities") pod "4f46f2e2-93c0-49a3-b901-3bd9e06f6342" (UID: "4f46f2e2-93c0-49a3-b901-3bd9e06f6342"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 17:08:16 crc kubenswrapper[4726]: I1001 17:08:16.865481 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 17:08:16 crc kubenswrapper[4726]: I1001 17:08:16.870446 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-kube-api-access-phlsj" (OuterVolumeSpecName: "kube-api-access-phlsj") pod "4f46f2e2-93c0-49a3-b901-3bd9e06f6342" (UID: "4f46f2e2-93c0-49a3-b901-3bd9e06f6342"). InnerVolumeSpecName "kube-api-access-phlsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 17:08:16 crc kubenswrapper[4726]: I1001 17:08:16.913406 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4f46f2e2-93c0-49a3-b901-3bd9e06f6342" (UID: "4f46f2e2-93c0-49a3-b901-3bd9e06f6342"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 17:08:16 crc kubenswrapper[4726]: I1001 17:08:16.967005 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 17:08:16 crc kubenswrapper[4726]: I1001 17:08:16.967074 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phlsj\" (UniqueName: \"kubernetes.io/projected/4f46f2e2-93c0-49a3-b901-3bd9e06f6342-kube-api-access-phlsj\") on node \"crc\" DevicePath \"\"" Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.300348 4726 generic.go:334] "Generic (PLEG): container finished" podID="4f46f2e2-93c0-49a3-b901-3bd9e06f6342" containerID="08bf037d93bf1c62583bb87d7f3c9616d25e86d83a86e4f213901670480b6a7f" exitCode=0 Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.300394 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49xhx" event={"ID":"4f46f2e2-93c0-49a3-b901-3bd9e06f6342","Type":"ContainerDied","Data":"08bf037d93bf1c62583bb87d7f3c9616d25e86d83a86e4f213901670480b6a7f"} Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.300422 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49xhx" Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.300436 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49xhx" event={"ID":"4f46f2e2-93c0-49a3-b901-3bd9e06f6342","Type":"ContainerDied","Data":"835fe2fc2f70d6e6243f0aa0f68c380ba29fc3fb1412ef2c434fb2e27c1054a4"} Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.300458 4726 scope.go:117] "RemoveContainer" containerID="08bf037d93bf1c62583bb87d7f3c9616d25e86d83a86e4f213901670480b6a7f" Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.322408 4726 scope.go:117] "RemoveContainer" containerID="f39287131613603a4273ea8139fe23ec6851aa1884b5023cf2545c48774fd571" Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.346001 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-49xhx"] Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.356872 4726 scope.go:117] "RemoveContainer" containerID="41966207ce8b2883fbdff899448e9ca3c40457e02017ac2d4bac3081022c4117" Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.358500 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-49xhx"] Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.390139 4726 scope.go:117] "RemoveContainer" containerID="08bf037d93bf1c62583bb87d7f3c9616d25e86d83a86e4f213901670480b6a7f" Oct 01 17:08:17 crc kubenswrapper[4726]: E1001 17:08:17.390509 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08bf037d93bf1c62583bb87d7f3c9616d25e86d83a86e4f213901670480b6a7f\": container with ID starting with 08bf037d93bf1c62583bb87d7f3c9616d25e86d83a86e4f213901670480b6a7f not found: ID does not exist" containerID="08bf037d93bf1c62583bb87d7f3c9616d25e86d83a86e4f213901670480b6a7f" Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.390555 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08bf037d93bf1c62583bb87d7f3c9616d25e86d83a86e4f213901670480b6a7f"} err="failed to get container status \"08bf037d93bf1c62583bb87d7f3c9616d25e86d83a86e4f213901670480b6a7f\": rpc error: code = NotFound desc = could not find container \"08bf037d93bf1c62583bb87d7f3c9616d25e86d83a86e4f213901670480b6a7f\": container with ID starting with 08bf037d93bf1c62583bb87d7f3c9616d25e86d83a86e4f213901670480b6a7f not found: ID does not exist" Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.390586 4726 scope.go:117] "RemoveContainer" containerID="f39287131613603a4273ea8139fe23ec6851aa1884b5023cf2545c48774fd571" Oct 01 17:08:17 crc kubenswrapper[4726]: E1001 17:08:17.390864 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f39287131613603a4273ea8139fe23ec6851aa1884b5023cf2545c48774fd571\": container with ID starting with f39287131613603a4273ea8139fe23ec6851aa1884b5023cf2545c48774fd571 not found: ID does not exist" containerID="f39287131613603a4273ea8139fe23ec6851aa1884b5023cf2545c48774fd571" Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.390890 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f39287131613603a4273ea8139fe23ec6851aa1884b5023cf2545c48774fd571"} err="failed to get container status \"f39287131613603a4273ea8139fe23ec6851aa1884b5023cf2545c48774fd571\": rpc error: code = NotFound desc = could not find container \"f39287131613603a4273ea8139fe23ec6851aa1884b5023cf2545c48774fd571\": container with ID starting with f39287131613603a4273ea8139fe23ec6851aa1884b5023cf2545c48774fd571 not found: ID does not exist" Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.390908 4726 scope.go:117] "RemoveContainer" containerID="41966207ce8b2883fbdff899448e9ca3c40457e02017ac2d4bac3081022c4117" Oct 01 17:08:17 crc kubenswrapper[4726]: E1001 17:08:17.391287 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41966207ce8b2883fbdff899448e9ca3c40457e02017ac2d4bac3081022c4117\": container with ID starting with 41966207ce8b2883fbdff899448e9ca3c40457e02017ac2d4bac3081022c4117 not found: ID does not exist" containerID="41966207ce8b2883fbdff899448e9ca3c40457e02017ac2d4bac3081022c4117" Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.391306 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41966207ce8b2883fbdff899448e9ca3c40457e02017ac2d4bac3081022c4117"} err="failed to get container status \"41966207ce8b2883fbdff899448e9ca3c40457e02017ac2d4bac3081022c4117\": rpc error: code = NotFound desc = could not find container \"41966207ce8b2883fbdff899448e9ca3c40457e02017ac2d4bac3081022c4117\": container with ID starting with 41966207ce8b2883fbdff899448e9ca3c40457e02017ac2d4bac3081022c4117 not found: ID does not exist" Oct 01 17:08:17 crc kubenswrapper[4726]: I1001 17:08:17.817190 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f46f2e2-93c0-49a3-b901-3bd9e06f6342" path="/var/lib/kubelet/pods/4f46f2e2-93c0-49a3-b901-3bd9e06f6342/volumes" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.413236 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.413841 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.413918 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.414917 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6a39a92e4a63907159963d04757f77099dc1afa2ff471236819de46bcfbf3573"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.414995 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://6a39a92e4a63907159963d04757f77099dc1afa2ff471236819de46bcfbf3573" gracePeriod=600 Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.471528 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bbw7f"] Oct 01 17:08:23 crc kubenswrapper[4726]: E1001 17:08:23.472384 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f46f2e2-93c0-49a3-b901-3bd9e06f6342" containerName="extract-utilities" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.472409 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f46f2e2-93c0-49a3-b901-3bd9e06f6342" containerName="extract-utilities" Oct 01 17:08:23 crc kubenswrapper[4726]: E1001 17:08:23.472439 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f46f2e2-93c0-49a3-b901-3bd9e06f6342" containerName="extract-content" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.472449 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f46f2e2-93c0-49a3-b901-3bd9e06f6342" containerName="extract-content" Oct 01 17:08:23 crc kubenswrapper[4726]: E1001 17:08:23.472460 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f46f2e2-93c0-49a3-b901-3bd9e06f6342" containerName="registry-server" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.472468 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f46f2e2-93c0-49a3-b901-3bd9e06f6342" containerName="registry-server" Oct 01 17:08:23 crc kubenswrapper[4726]: E1001 17:08:23.472491 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edca44cc-3040-4598-9e7c-ed7a7b235771" containerName="container-00" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.472498 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="edca44cc-3040-4598-9e7c-ed7a7b235771" containerName="container-00" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.472723 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="edca44cc-3040-4598-9e7c-ed7a7b235771" containerName="container-00" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.472757 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f46f2e2-93c0-49a3-b901-3bd9e06f6342" containerName="registry-server" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.474435 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.484686 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbw7f"] Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.528558 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rh5j\" (UniqueName: \"kubernetes.io/projected/aa560228-7ccf-4aa9-aa88-2ac78de2f172-kube-api-access-7rh5j\") pod \"redhat-marketplace-bbw7f\" (UID: \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\") " pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.528620 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa560228-7ccf-4aa9-aa88-2ac78de2f172-utilities\") pod \"redhat-marketplace-bbw7f\" (UID: \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\") " pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.528641 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa560228-7ccf-4aa9-aa88-2ac78de2f172-catalog-content\") pod \"redhat-marketplace-bbw7f\" (UID: \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\") " pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.630236 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rh5j\" (UniqueName: \"kubernetes.io/projected/aa560228-7ccf-4aa9-aa88-2ac78de2f172-kube-api-access-7rh5j\") pod \"redhat-marketplace-bbw7f\" (UID: \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\") " pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.630310 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa560228-7ccf-4aa9-aa88-2ac78de2f172-utilities\") pod \"redhat-marketplace-bbw7f\" (UID: \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\") " pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.630350 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa560228-7ccf-4aa9-aa88-2ac78de2f172-catalog-content\") pod \"redhat-marketplace-bbw7f\" (UID: \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\") " pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.631086 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa560228-7ccf-4aa9-aa88-2ac78de2f172-utilities\") pod \"redhat-marketplace-bbw7f\" (UID: \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\") " pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.631095 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa560228-7ccf-4aa9-aa88-2ac78de2f172-catalog-content\") pod \"redhat-marketplace-bbw7f\" (UID: \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\") " pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.654355 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rh5j\" (UniqueName: \"kubernetes.io/projected/aa560228-7ccf-4aa9-aa88-2ac78de2f172-kube-api-access-7rh5j\") pod \"redhat-marketplace-bbw7f\" (UID: \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\") " pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:23 crc kubenswrapper[4726]: I1001 17:08:23.876985 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:24 crc kubenswrapper[4726]: I1001 17:08:24.313045 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbw7f"] Oct 01 17:08:24 crc kubenswrapper[4726]: W1001 17:08:24.318740 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa560228_7ccf_4aa9_aa88_2ac78de2f172.slice/crio-0fb7e9b8f177962a8a23607374ebd84c9e439d7e048d62eb1d017def8d384aa1 WatchSource:0}: Error finding container 0fb7e9b8f177962a8a23607374ebd84c9e439d7e048d62eb1d017def8d384aa1: Status 404 returned error can't find the container with id 0fb7e9b8f177962a8a23607374ebd84c9e439d7e048d62eb1d017def8d384aa1 Oct 01 17:08:24 crc kubenswrapper[4726]: I1001 17:08:24.388399 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbw7f" event={"ID":"aa560228-7ccf-4aa9-aa88-2ac78de2f172","Type":"ContainerStarted","Data":"0fb7e9b8f177962a8a23607374ebd84c9e439d7e048d62eb1d017def8d384aa1"} Oct 01 17:08:24 crc kubenswrapper[4726]: I1001 17:08:24.390718 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="6a39a92e4a63907159963d04757f77099dc1afa2ff471236819de46bcfbf3573" exitCode=0 Oct 01 17:08:24 crc kubenswrapper[4726]: I1001 17:08:24.390764 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"6a39a92e4a63907159963d04757f77099dc1afa2ff471236819de46bcfbf3573"} Oct 01 17:08:24 crc kubenswrapper[4726]: I1001 17:08:24.390797 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerStarted","Data":"f636fea95e8fc2020d3ea5c194d521a170b6d8d9b7ff2acf2fbaba2ddad73ed8"} Oct 01 17:08:24 crc kubenswrapper[4726]: I1001 17:08:24.390817 4726 scope.go:117] "RemoveContainer" containerID="8a29319da67f1df97e3611ad9264f38033c3ae03c23f1d4feb2bb76840abb865" Oct 01 17:08:25 crc kubenswrapper[4726]: I1001 17:08:25.406622 4726 generic.go:334] "Generic (PLEG): container finished" podID="aa560228-7ccf-4aa9-aa88-2ac78de2f172" containerID="34c97bfbde38a0ecf39bbcfa01749346c9fc491909b0cb4170ffa4e5a5ad3af6" exitCode=0 Oct 01 17:08:25 crc kubenswrapper[4726]: I1001 17:08:25.406756 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbw7f" event={"ID":"aa560228-7ccf-4aa9-aa88-2ac78de2f172","Type":"ContainerDied","Data":"34c97bfbde38a0ecf39bbcfa01749346c9fc491909b0cb4170ffa4e5a5ad3af6"} Oct 01 17:08:25 crc kubenswrapper[4726]: I1001 17:08:25.412299 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 17:08:25 crc kubenswrapper[4726]: I1001 17:08:25.570741 4726 scope.go:117] "RemoveContainer" containerID="5aecefd230d7bcfd7df54874dd7991638bf56d76689302a7882ca1c83ef17e13" Oct 01 17:08:27 crc kubenswrapper[4726]: I1001 17:08:27.435650 4726 generic.go:334] "Generic (PLEG): container finished" podID="aa560228-7ccf-4aa9-aa88-2ac78de2f172" containerID="38f0ac010bda1c9304892838ed3db15544ed02a0bcf55e2fd3c2c7757eb58af9" exitCode=0 Oct 01 17:08:27 crc kubenswrapper[4726]: I1001 17:08:27.435778 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbw7f" event={"ID":"aa560228-7ccf-4aa9-aa88-2ac78de2f172","Type":"ContainerDied","Data":"38f0ac010bda1c9304892838ed3db15544ed02a0bcf55e2fd3c2c7757eb58af9"} Oct 01 17:08:28 crc kubenswrapper[4726]: I1001 17:08:28.454923 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbw7f" event={"ID":"aa560228-7ccf-4aa9-aa88-2ac78de2f172","Type":"ContainerStarted","Data":"af599926094014e4dac6f1483f30d0c9d12ec8bc93cbbe84ae15d542d6f49a06"} Oct 01 17:08:28 crc kubenswrapper[4726]: I1001 17:08:28.489490 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bbw7f" podStartSLOduration=2.9819877 podStartE2EDuration="5.489462378s" podCreationTimestamp="2025-10-01 17:08:23 +0000 UTC" firstStartedPulling="2025-10-01 17:08:25.411697875 +0000 UTC m=+4158.313250502" lastFinishedPulling="2025-10-01 17:08:27.919172603 +0000 UTC m=+4160.820725180" observedRunningTime="2025-10-01 17:08:28.481904992 +0000 UTC m=+4161.383457569" watchObservedRunningTime="2025-10-01 17:08:28.489462378 +0000 UTC m=+4161.391014995" Oct 01 17:08:29 crc kubenswrapper[4726]: I1001 17:08:29.247169 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-z4ml2_8282f47b-cd59-4cba-a1aa-7528d57842ec/control-plane-machine-set-operator/0.log" Oct 01 17:08:29 crc kubenswrapper[4726]: I1001 17:08:29.426003 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mpfpx_5399581d-2111-47d1-bfe7-007b7c46d706/kube-rbac-proxy/0.log" Oct 01 17:08:29 crc kubenswrapper[4726]: I1001 17:08:29.499668 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mpfpx_5399581d-2111-47d1-bfe7-007b7c46d706/machine-api-operator/0.log" Oct 01 17:08:33 crc kubenswrapper[4726]: I1001 17:08:33.878023 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:33 crc kubenswrapper[4726]: I1001 17:08:33.878566 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:33 crc kubenswrapper[4726]: I1001 17:08:33.945137 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:34 crc kubenswrapper[4726]: I1001 17:08:34.568433 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:34 crc kubenswrapper[4726]: I1001 17:08:34.898351 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbw7f"] Oct 01 17:08:36 crc kubenswrapper[4726]: I1001 17:08:36.535007 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bbw7f" podUID="aa560228-7ccf-4aa9-aa88-2ac78de2f172" containerName="registry-server" containerID="cri-o://af599926094014e4dac6f1483f30d0c9d12ec8bc93cbbe84ae15d542d6f49a06" gracePeriod=2 Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.012135 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.096709 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa560228-7ccf-4aa9-aa88-2ac78de2f172-catalog-content\") pod \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\" (UID: \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\") " Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.096789 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa560228-7ccf-4aa9-aa88-2ac78de2f172-utilities\") pod \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\" (UID: \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\") " Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.096871 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rh5j\" (UniqueName: \"kubernetes.io/projected/aa560228-7ccf-4aa9-aa88-2ac78de2f172-kube-api-access-7rh5j\") pod \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\" (UID: \"aa560228-7ccf-4aa9-aa88-2ac78de2f172\") " Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.097689 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa560228-7ccf-4aa9-aa88-2ac78de2f172-utilities" (OuterVolumeSpecName: "utilities") pod "aa560228-7ccf-4aa9-aa88-2ac78de2f172" (UID: "aa560228-7ccf-4aa9-aa88-2ac78de2f172"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.106370 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa560228-7ccf-4aa9-aa88-2ac78de2f172-kube-api-access-7rh5j" (OuterVolumeSpecName: "kube-api-access-7rh5j") pod "aa560228-7ccf-4aa9-aa88-2ac78de2f172" (UID: "aa560228-7ccf-4aa9-aa88-2ac78de2f172"). InnerVolumeSpecName "kube-api-access-7rh5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.110558 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa560228-7ccf-4aa9-aa88-2ac78de2f172-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa560228-7ccf-4aa9-aa88-2ac78de2f172" (UID: "aa560228-7ccf-4aa9-aa88-2ac78de2f172"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.199781 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa560228-7ccf-4aa9-aa88-2ac78de2f172-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.199814 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa560228-7ccf-4aa9-aa88-2ac78de2f172-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.199825 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rh5j\" (UniqueName: \"kubernetes.io/projected/aa560228-7ccf-4aa9-aa88-2ac78de2f172-kube-api-access-7rh5j\") on node \"crc\" DevicePath \"\"" Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.549997 4726 generic.go:334] "Generic (PLEG): container finished" podID="aa560228-7ccf-4aa9-aa88-2ac78de2f172" containerID="af599926094014e4dac6f1483f30d0c9d12ec8bc93cbbe84ae15d542d6f49a06" exitCode=0 Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.550100 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbw7f" event={"ID":"aa560228-7ccf-4aa9-aa88-2ac78de2f172","Type":"ContainerDied","Data":"af599926094014e4dac6f1483f30d0c9d12ec8bc93cbbe84ae15d542d6f49a06"} Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.550467 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbw7f" event={"ID":"aa560228-7ccf-4aa9-aa88-2ac78de2f172","Type":"ContainerDied","Data":"0fb7e9b8f177962a8a23607374ebd84c9e439d7e048d62eb1d017def8d384aa1"} Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.550497 4726 scope.go:117] "RemoveContainer" containerID="af599926094014e4dac6f1483f30d0c9d12ec8bc93cbbe84ae15d542d6f49a06" Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.550227 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bbw7f" Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.577703 4726 scope.go:117] "RemoveContainer" containerID="38f0ac010bda1c9304892838ed3db15544ed02a0bcf55e2fd3c2c7757eb58af9" Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.626665 4726 scope.go:117] "RemoveContainer" containerID="34c97bfbde38a0ecf39bbcfa01749346c9fc491909b0cb4170ffa4e5a5ad3af6" Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.627733 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbw7f"] Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.637265 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbw7f"] Oct 01 17:08:37 crc kubenswrapper[4726]: I1001 17:08:37.821881 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa560228-7ccf-4aa9-aa88-2ac78de2f172" path="/var/lib/kubelet/pods/aa560228-7ccf-4aa9-aa88-2ac78de2f172/volumes" Oct 01 17:08:38 crc kubenswrapper[4726]: I1001 17:08:38.326558 4726 scope.go:117] "RemoveContainer" containerID="af599926094014e4dac6f1483f30d0c9d12ec8bc93cbbe84ae15d542d6f49a06" Oct 01 17:08:38 crc kubenswrapper[4726]: E1001 17:08:38.327159 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af599926094014e4dac6f1483f30d0c9d12ec8bc93cbbe84ae15d542d6f49a06\": container with ID starting with af599926094014e4dac6f1483f30d0c9d12ec8bc93cbbe84ae15d542d6f49a06 not found: ID does not exist" containerID="af599926094014e4dac6f1483f30d0c9d12ec8bc93cbbe84ae15d542d6f49a06" Oct 01 17:08:38 crc kubenswrapper[4726]: I1001 17:08:38.327199 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af599926094014e4dac6f1483f30d0c9d12ec8bc93cbbe84ae15d542d6f49a06"} err="failed to get container status \"af599926094014e4dac6f1483f30d0c9d12ec8bc93cbbe84ae15d542d6f49a06\": rpc error: code = NotFound desc = could not find container \"af599926094014e4dac6f1483f30d0c9d12ec8bc93cbbe84ae15d542d6f49a06\": container with ID starting with af599926094014e4dac6f1483f30d0c9d12ec8bc93cbbe84ae15d542d6f49a06 not found: ID does not exist" Oct 01 17:08:38 crc kubenswrapper[4726]: I1001 17:08:38.327228 4726 scope.go:117] "RemoveContainer" containerID="38f0ac010bda1c9304892838ed3db15544ed02a0bcf55e2fd3c2c7757eb58af9" Oct 01 17:08:38 crc kubenswrapper[4726]: E1001 17:08:38.327620 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38f0ac010bda1c9304892838ed3db15544ed02a0bcf55e2fd3c2c7757eb58af9\": container with ID starting with 38f0ac010bda1c9304892838ed3db15544ed02a0bcf55e2fd3c2c7757eb58af9 not found: ID does not exist" containerID="38f0ac010bda1c9304892838ed3db15544ed02a0bcf55e2fd3c2c7757eb58af9" Oct 01 17:08:38 crc kubenswrapper[4726]: I1001 17:08:38.327648 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f0ac010bda1c9304892838ed3db15544ed02a0bcf55e2fd3c2c7757eb58af9"} err="failed to get container status \"38f0ac010bda1c9304892838ed3db15544ed02a0bcf55e2fd3c2c7757eb58af9\": rpc error: code = NotFound desc = could not find container \"38f0ac010bda1c9304892838ed3db15544ed02a0bcf55e2fd3c2c7757eb58af9\": container with ID starting with 38f0ac010bda1c9304892838ed3db15544ed02a0bcf55e2fd3c2c7757eb58af9 not found: ID does not exist" Oct 01 17:08:38 crc kubenswrapper[4726]: I1001 17:08:38.327665 4726 scope.go:117] "RemoveContainer" containerID="34c97bfbde38a0ecf39bbcfa01749346c9fc491909b0cb4170ffa4e5a5ad3af6" Oct 01 17:08:38 crc kubenswrapper[4726]: E1001 17:08:38.327991 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34c97bfbde38a0ecf39bbcfa01749346c9fc491909b0cb4170ffa4e5a5ad3af6\": container with ID starting with 34c97bfbde38a0ecf39bbcfa01749346c9fc491909b0cb4170ffa4e5a5ad3af6 not found: ID does not exist" containerID="34c97bfbde38a0ecf39bbcfa01749346c9fc491909b0cb4170ffa4e5a5ad3af6" Oct 01 17:08:38 crc kubenswrapper[4726]: I1001 17:08:38.328021 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34c97bfbde38a0ecf39bbcfa01749346c9fc491909b0cb4170ffa4e5a5ad3af6"} err="failed to get container status \"34c97bfbde38a0ecf39bbcfa01749346c9fc491909b0cb4170ffa4e5a5ad3af6\": rpc error: code = NotFound desc = could not find container \"34c97bfbde38a0ecf39bbcfa01749346c9fc491909b0cb4170ffa4e5a5ad3af6\": container with ID starting with 34c97bfbde38a0ecf39bbcfa01749346c9fc491909b0cb4170ffa4e5a5ad3af6 not found: ID does not exist" Oct 01 17:08:41 crc kubenswrapper[4726]: I1001 17:08:41.444275 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-jkfpv_cfc89425-6bca-4267-9a51-50e0d8325eaa/cert-manager-controller/0.log" Oct 01 17:08:41 crc kubenswrapper[4726]: I1001 17:08:41.576336 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-pz8pb_55df3602-b3df-45f4-90b8-eddb1903bf32/cert-manager-cainjector/0.log" Oct 01 17:08:41 crc kubenswrapper[4726]: I1001 17:08:41.633017 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-jxd7m_51706d0c-886c-4865-ba8b-ade37f39b84e/cert-manager-webhook/0.log" Oct 01 17:08:53 crc kubenswrapper[4726]: I1001 17:08:53.502190 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-vqzfm_5fb896f2-8fbc-40d7-adf4-2a93089171ce/nmstate-console-plugin/0.log" Oct 01 17:08:53 crc kubenswrapper[4726]: I1001 17:08:53.703866 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-kzwz4_fc59e6f9-76eb-49fb-9028-9e64820de406/nmstate-handler/0.log" Oct 01 17:08:53 crc kubenswrapper[4726]: I1001 17:08:53.765564 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-xxpvz_d24b198c-db7b-4057-bed1-6f32d0bac5d3/nmstate-metrics/0.log" Oct 01 17:08:53 crc kubenswrapper[4726]: I1001 17:08:53.820843 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-xxpvz_d24b198c-db7b-4057-bed1-6f32d0bac5d3/kube-rbac-proxy/0.log" Oct 01 17:08:53 crc kubenswrapper[4726]: I1001 17:08:53.948313 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-mn7jt_d4e4c724-1025-48b6-a04c-77071be69269/nmstate-operator/0.log" Oct 01 17:08:53 crc kubenswrapper[4726]: I1001 17:08:53.993369 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-qrkz2_0f80c25b-98a7-46d0-bf07-26c19f885781/nmstate-webhook/0.log" Oct 01 17:09:08 crc kubenswrapper[4726]: I1001 17:09:08.148796 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-kk7mh_1c256ab6-be12-4843-a072-f6bb861c9740/kube-rbac-proxy/0.log" Oct 01 17:09:08 crc kubenswrapper[4726]: I1001 17:09:08.223701 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-kk7mh_1c256ab6-be12-4843-a072-f6bb861c9740/controller/0.log" Oct 01 17:09:08 crc kubenswrapper[4726]: I1001 17:09:08.307220 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-mqdn2_678af4b7-7942-40da-8273-de583fd22666/frr-k8s-webhook-server/0.log" Oct 01 17:09:08 crc kubenswrapper[4726]: I1001 17:09:08.429080 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-frr-files/0.log" Oct 01 17:09:08 crc kubenswrapper[4726]: I1001 17:09:08.565749 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-metrics/0.log" Oct 01 17:09:08 crc kubenswrapper[4726]: I1001 17:09:08.576038 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-reloader/0.log" Oct 01 17:09:08 crc kubenswrapper[4726]: I1001 17:09:08.604293 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-frr-files/0.log" Oct 01 17:09:08 crc kubenswrapper[4726]: I1001 17:09:08.634927 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-reloader/0.log" Oct 01 17:09:08 crc kubenswrapper[4726]: I1001 17:09:08.825333 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-frr-files/0.log" Oct 01 17:09:08 crc kubenswrapper[4726]: I1001 17:09:08.853466 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-metrics/0.log" Oct 01 17:09:08 crc kubenswrapper[4726]: I1001 17:09:08.863455 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-reloader/0.log" Oct 01 17:09:08 crc kubenswrapper[4726]: I1001 17:09:08.884530 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-metrics/0.log" Oct 01 17:09:09 crc kubenswrapper[4726]: I1001 17:09:09.055375 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-frr-files/0.log" Oct 01 17:09:09 crc kubenswrapper[4726]: I1001 17:09:09.056266 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-reloader/0.log" Oct 01 17:09:09 crc kubenswrapper[4726]: I1001 17:09:09.080441 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/controller/0.log" Oct 01 17:09:09 crc kubenswrapper[4726]: I1001 17:09:09.086748 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/cp-metrics/0.log" Oct 01 17:09:09 crc kubenswrapper[4726]: I1001 17:09:09.227303 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/frr-metrics/0.log" Oct 01 17:09:09 crc kubenswrapper[4726]: I1001 17:09:09.284350 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/kube-rbac-proxy/0.log" Oct 01 17:09:09 crc kubenswrapper[4726]: I1001 17:09:09.295845 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/kube-rbac-proxy-frr/0.log" Oct 01 17:09:09 crc kubenswrapper[4726]: I1001 17:09:09.589829 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/frr/0.log" Oct 01 17:09:09 crc kubenswrapper[4726]: I1001 17:09:09.784273 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xqs68_9b9a95f4-3ea1-45a0-8884-b6dc26ad5895/reloader/0.log" Oct 01 17:09:09 crc kubenswrapper[4726]: I1001 17:09:09.787813 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5fc7dbb978-2m6rr_9dd491a2-a155-401a-86fb-04e5fe23984e/manager/0.log" Oct 01 17:09:09 crc kubenswrapper[4726]: I1001 17:09:09.978555 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-d8cc586f4-l28zk_f729348c-0bd7-4667-86f9-94e36a9afb11/webhook-server/0.log" Oct 01 17:09:10 crc kubenswrapper[4726]: I1001 17:09:10.005735 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tdc4_087620f7-11b5-49b6-a8b7-30a75a5196cb/kube-rbac-proxy/0.log" Oct 01 17:09:10 crc kubenswrapper[4726]: I1001 17:09:10.139899 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tdc4_087620f7-11b5-49b6-a8b7-30a75a5196cb/speaker/0.log" Oct 01 17:09:23 crc kubenswrapper[4726]: I1001 17:09:23.553918 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2_7b3a32c8-9c47-4ccc-9885-d29009a22aa4/util/0.log" Oct 01 17:09:23 crc kubenswrapper[4726]: I1001 17:09:23.739813 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2_7b3a32c8-9c47-4ccc-9885-d29009a22aa4/pull/0.log" Oct 01 17:09:23 crc kubenswrapper[4726]: I1001 17:09:23.775111 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2_7b3a32c8-9c47-4ccc-9885-d29009a22aa4/util/0.log" Oct 01 17:09:23 crc kubenswrapper[4726]: I1001 17:09:23.789643 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2_7b3a32c8-9c47-4ccc-9885-d29009a22aa4/pull/0.log" Oct 01 17:09:23 crc kubenswrapper[4726]: I1001 17:09:23.976161 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2_7b3a32c8-9c47-4ccc-9885-d29009a22aa4/util/0.log" Oct 01 17:09:23 crc kubenswrapper[4726]: I1001 17:09:23.980019 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2_7b3a32c8-9c47-4ccc-9885-d29009a22aa4/extract/0.log" Oct 01 17:09:24 crc kubenswrapper[4726]: I1001 17:09:24.029676 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22kdl2_7b3a32c8-9c47-4ccc-9885-d29009a22aa4/pull/0.log" Oct 01 17:09:24 crc kubenswrapper[4726]: I1001 17:09:24.172769 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_f2e6c6e1-9224-4c26-adcf-07fdf1ea4538/util/0.log" Oct 01 17:09:24 crc kubenswrapper[4726]: I1001 17:09:24.348597 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_f2e6c6e1-9224-4c26-adcf-07fdf1ea4538/pull/0.log" Oct 01 17:09:24 crc kubenswrapper[4726]: I1001 17:09:24.372501 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_f2e6c6e1-9224-4c26-adcf-07fdf1ea4538/pull/0.log" Oct 01 17:09:24 crc kubenswrapper[4726]: I1001 17:09:24.425797 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_f2e6c6e1-9224-4c26-adcf-07fdf1ea4538/util/0.log" Oct 01 17:09:24 crc kubenswrapper[4726]: I1001 17:09:24.553119 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_f2e6c6e1-9224-4c26-adcf-07fdf1ea4538/pull/0.log" Oct 01 17:09:24 crc kubenswrapper[4726]: I1001 17:09:24.588434 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_f2e6c6e1-9224-4c26-adcf-07fdf1ea4538/util/0.log" Oct 01 17:09:24 crc kubenswrapper[4726]: I1001 17:09:24.643875 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc2dnns_f2e6c6e1-9224-4c26-adcf-07fdf1ea4538/extract/0.log" Oct 01 17:09:24 crc kubenswrapper[4726]: I1001 17:09:24.784452 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q5bzx_9f2ddc53-446b-42b0-966a-f6b2f4eb6159/extract-utilities/0.log" Oct 01 17:09:24 crc kubenswrapper[4726]: I1001 17:09:24.984006 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q5bzx_9f2ddc53-446b-42b0-966a-f6b2f4eb6159/extract-content/0.log" Oct 01 17:09:24 crc kubenswrapper[4726]: I1001 17:09:24.994922 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q5bzx_9f2ddc53-446b-42b0-966a-f6b2f4eb6159/extract-utilities/0.log" Oct 01 17:09:25 crc kubenswrapper[4726]: I1001 17:09:25.007247 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q5bzx_9f2ddc53-446b-42b0-966a-f6b2f4eb6159/extract-content/0.log" Oct 01 17:09:25 crc kubenswrapper[4726]: I1001 17:09:25.362892 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q5bzx_9f2ddc53-446b-42b0-966a-f6b2f4eb6159/extract-content/0.log" Oct 01 17:09:25 crc kubenswrapper[4726]: I1001 17:09:25.415705 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q5bzx_9f2ddc53-446b-42b0-966a-f6b2f4eb6159/extract-utilities/0.log" Oct 01 17:09:25 crc kubenswrapper[4726]: I1001 17:09:25.543702 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ch575_039da856-2b47-4aff-b18b-9e9cff090200/extract-utilities/0.log" Oct 01 17:09:25 crc kubenswrapper[4726]: I1001 17:09:25.693511 4726 scope.go:117] "RemoveContainer" containerID="f4dc498758b1860d5b1149f025950f8ece0663b1536add46dcc7ed295a7515c7" Oct 01 17:09:25 crc kubenswrapper[4726]: I1001 17:09:25.719646 4726 scope.go:117] "RemoveContainer" containerID="cc97ee47dfcedca58bdd09ec3dfd591f517b41ae60fe61a02e96ee055ab0f3df" Oct 01 17:09:25 crc kubenswrapper[4726]: I1001 17:09:25.767801 4726 scope.go:117] "RemoveContainer" containerID="9f550cc4eca532eb5554a6cc49b298744885a461d3fb92fdd15720bd05f1e87a" Oct 01 17:09:25 crc kubenswrapper[4726]: I1001 17:09:25.848363 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ch575_039da856-2b47-4aff-b18b-9e9cff090200/extract-utilities/0.log" Oct 01 17:09:25 crc kubenswrapper[4726]: I1001 17:09:25.866261 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ch575_039da856-2b47-4aff-b18b-9e9cff090200/extract-content/0.log" Oct 01 17:09:25 crc kubenswrapper[4726]: I1001 17:09:25.883668 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q5bzx_9f2ddc53-446b-42b0-966a-f6b2f4eb6159/registry-server/0.log" Oct 01 17:09:25 crc kubenswrapper[4726]: I1001 17:09:25.910512 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ch575_039da856-2b47-4aff-b18b-9e9cff090200/extract-content/0.log" Oct 01 17:09:26 crc kubenswrapper[4726]: I1001 17:09:26.076117 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ch575_039da856-2b47-4aff-b18b-9e9cff090200/extract-content/0.log" Oct 01 17:09:26 crc kubenswrapper[4726]: I1001 17:09:26.080450 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ch575_039da856-2b47-4aff-b18b-9e9cff090200/extract-utilities/0.log" Oct 01 17:09:26 crc kubenswrapper[4726]: I1001 17:09:26.307778 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b_b751f5f8-ea87-4fcb-8c07-93863ae71d03/util/0.log" Oct 01 17:09:26 crc kubenswrapper[4726]: I1001 17:09:26.586319 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b_b751f5f8-ea87-4fcb-8c07-93863ae71d03/util/0.log" Oct 01 17:09:26 crc kubenswrapper[4726]: I1001 17:09:26.598697 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b_b751f5f8-ea87-4fcb-8c07-93863ae71d03/pull/0.log" Oct 01 17:09:26 crc kubenswrapper[4726]: I1001 17:09:26.653737 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b_b751f5f8-ea87-4fcb-8c07-93863ae71d03/pull/0.log" Oct 01 17:09:26 crc kubenswrapper[4726]: I1001 17:09:26.704847 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ch575_039da856-2b47-4aff-b18b-9e9cff090200/registry-server/0.log" Oct 01 17:09:26 crc kubenswrapper[4726]: I1001 17:09:26.780718 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b_b751f5f8-ea87-4fcb-8c07-93863ae71d03/util/0.log" Oct 01 17:09:26 crc kubenswrapper[4726]: I1001 17:09:26.791260 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b_b751f5f8-ea87-4fcb-8c07-93863ae71d03/extract/0.log" Oct 01 17:09:26 crc kubenswrapper[4726]: I1001 17:09:26.803804 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96wf58b_b751f5f8-ea87-4fcb-8c07-93863ae71d03/pull/0.log" Oct 01 17:09:26 crc kubenswrapper[4726]: I1001 17:09:26.888778 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s_213c0beb-44a8-4e18-afaa-4d2ba6fd6305/util/0.log" Oct 01 17:09:27 crc kubenswrapper[4726]: I1001 17:09:27.101190 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s_213c0beb-44a8-4e18-afaa-4d2ba6fd6305/pull/0.log" Oct 01 17:09:27 crc kubenswrapper[4726]: I1001 17:09:27.125340 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s_213c0beb-44a8-4e18-afaa-4d2ba6fd6305/pull/0.log" Oct 01 17:09:27 crc kubenswrapper[4726]: I1001 17:09:27.131818 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s_213c0beb-44a8-4e18-afaa-4d2ba6fd6305/util/0.log" Oct 01 17:09:27 crc kubenswrapper[4726]: I1001 17:09:27.474571 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s_213c0beb-44a8-4e18-afaa-4d2ba6fd6305/pull/0.log" Oct 01 17:09:27 crc kubenswrapper[4726]: I1001 17:09:27.480588 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s_213c0beb-44a8-4e18-afaa-4d2ba6fd6305/extract/0.log" Oct 01 17:09:27 crc kubenswrapper[4726]: I1001 17:09:27.506487 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c2vb6s_213c0beb-44a8-4e18-afaa-4d2ba6fd6305/util/0.log" Oct 01 17:09:27 crc kubenswrapper[4726]: I1001 17:09:27.538043 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-g69q5_c460d271-b44c-482b-a0d9-70bdc879cf1d/marketplace-operator/0.log" Oct 01 17:09:27 crc kubenswrapper[4726]: I1001 17:09:27.680390 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ms6sz_4753c482-c4a6-44e9-ace2-7997ee15d241/extract-utilities/0.log" Oct 01 17:09:27 crc kubenswrapper[4726]: I1001 17:09:27.824586 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ms6sz_4753c482-c4a6-44e9-ace2-7997ee15d241/extract-utilities/0.log" Oct 01 17:09:27 crc kubenswrapper[4726]: I1001 17:09:27.842244 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ms6sz_4753c482-c4a6-44e9-ace2-7997ee15d241/extract-content/0.log" Oct 01 17:09:27 crc kubenswrapper[4726]: I1001 17:09:27.883040 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ms6sz_4753c482-c4a6-44e9-ace2-7997ee15d241/extract-content/0.log" Oct 01 17:09:28 crc kubenswrapper[4726]: I1001 17:09:28.087645 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nj4wf_c24a10a9-6d60-43d5-9e3c-67cec3e0045c/extract-utilities/0.log" Oct 01 17:09:28 crc kubenswrapper[4726]: I1001 17:09:28.096196 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ms6sz_4753c482-c4a6-44e9-ace2-7997ee15d241/extract-utilities/0.log" Oct 01 17:09:28 crc kubenswrapper[4726]: I1001 17:09:28.128480 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ms6sz_4753c482-c4a6-44e9-ace2-7997ee15d241/extract-content/0.log" Oct 01 17:09:28 crc kubenswrapper[4726]: I1001 17:09:28.323375 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ms6sz_4753c482-c4a6-44e9-ace2-7997ee15d241/registry-server/0.log" Oct 01 17:09:28 crc kubenswrapper[4726]: I1001 17:09:28.388499 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nj4wf_c24a10a9-6d60-43d5-9e3c-67cec3e0045c/extract-utilities/0.log" Oct 01 17:09:28 crc kubenswrapper[4726]: I1001 17:09:28.404921 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nj4wf_c24a10a9-6d60-43d5-9e3c-67cec3e0045c/extract-content/0.log" Oct 01 17:09:28 crc kubenswrapper[4726]: I1001 17:09:28.465649 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nj4wf_c24a10a9-6d60-43d5-9e3c-67cec3e0045c/extract-content/0.log" Oct 01 17:09:28 crc kubenswrapper[4726]: I1001 17:09:28.600548 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nj4wf_c24a10a9-6d60-43d5-9e3c-67cec3e0045c/extract-utilities/0.log" Oct 01 17:09:28 crc kubenswrapper[4726]: I1001 17:09:28.617420 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nj4wf_c24a10a9-6d60-43d5-9e3c-67cec3e0045c/extract-content/0.log" Oct 01 17:09:28 crc kubenswrapper[4726]: I1001 17:09:28.896282 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nj4wf_c24a10a9-6d60-43d5-9e3c-67cec3e0045c/registry-server/0.log" Oct 01 17:09:44 crc kubenswrapper[4726]: I1001 17:09:44.873023 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-97p87"] Oct 01 17:09:44 crc kubenswrapper[4726]: E1001 17:09:44.874209 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa560228-7ccf-4aa9-aa88-2ac78de2f172" containerName="extract-content" Oct 01 17:09:44 crc kubenswrapper[4726]: I1001 17:09:44.874229 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa560228-7ccf-4aa9-aa88-2ac78de2f172" containerName="extract-content" Oct 01 17:09:44 crc kubenswrapper[4726]: E1001 17:09:44.874258 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa560228-7ccf-4aa9-aa88-2ac78de2f172" containerName="registry-server" Oct 01 17:09:44 crc kubenswrapper[4726]: I1001 17:09:44.874266 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa560228-7ccf-4aa9-aa88-2ac78de2f172" containerName="registry-server" Oct 01 17:09:44 crc kubenswrapper[4726]: E1001 17:09:44.874287 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa560228-7ccf-4aa9-aa88-2ac78de2f172" containerName="extract-utilities" Oct 01 17:09:44 crc kubenswrapper[4726]: I1001 17:09:44.874296 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa560228-7ccf-4aa9-aa88-2ac78de2f172" containerName="extract-utilities" Oct 01 17:09:44 crc kubenswrapper[4726]: I1001 17:09:44.874535 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa560228-7ccf-4aa9-aa88-2ac78de2f172" containerName="registry-server" Oct 01 17:09:44 crc kubenswrapper[4726]: I1001 17:09:44.876639 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:44 crc kubenswrapper[4726]: I1001 17:09:44.887081 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-97p87"] Oct 01 17:09:44 crc kubenswrapper[4726]: I1001 17:09:44.977913 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-catalog-content\") pod \"community-operators-97p87\" (UID: \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\") " pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:44 crc kubenswrapper[4726]: I1001 17:09:44.978003 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-utilities\") pod \"community-operators-97p87\" (UID: \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\") " pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:44 crc kubenswrapper[4726]: I1001 17:09:44.978149 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdfg5\" (UniqueName: \"kubernetes.io/projected/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-kube-api-access-sdfg5\") pod \"community-operators-97p87\" (UID: \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\") " pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:45 crc kubenswrapper[4726]: I1001 17:09:45.080172 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-catalog-content\") pod \"community-operators-97p87\" (UID: \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\") " pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:45 crc kubenswrapper[4726]: I1001 17:09:45.080267 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-utilities\") pod \"community-operators-97p87\" (UID: \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\") " pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:45 crc kubenswrapper[4726]: I1001 17:09:45.080336 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdfg5\" (UniqueName: \"kubernetes.io/projected/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-kube-api-access-sdfg5\") pod \"community-operators-97p87\" (UID: \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\") " pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:45 crc kubenswrapper[4726]: I1001 17:09:45.080691 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-catalog-content\") pod \"community-operators-97p87\" (UID: \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\") " pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:45 crc kubenswrapper[4726]: I1001 17:09:45.080924 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-utilities\") pod \"community-operators-97p87\" (UID: \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\") " pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:45 crc kubenswrapper[4726]: I1001 17:09:45.109486 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdfg5\" (UniqueName: \"kubernetes.io/projected/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-kube-api-access-sdfg5\") pod \"community-operators-97p87\" (UID: \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\") " pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:45 crc kubenswrapper[4726]: I1001 17:09:45.203489 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:45 crc kubenswrapper[4726]: I1001 17:09:45.715490 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-97p87"] Oct 01 17:09:45 crc kubenswrapper[4726]: W1001 17:09:45.722594 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4407cffe_583e_4f8f_9da8_bc43ea8fcd32.slice/crio-774e1b4a49c20083b8bc356400f670652ea7588e8728541c1bb2cd3f0d794c34 WatchSource:0}: Error finding container 774e1b4a49c20083b8bc356400f670652ea7588e8728541c1bb2cd3f0d794c34: Status 404 returned error can't find the container with id 774e1b4a49c20083b8bc356400f670652ea7588e8728541c1bb2cd3f0d794c34 Oct 01 17:09:46 crc kubenswrapper[4726]: I1001 17:09:46.269223 4726 generic.go:334] "Generic (PLEG): container finished" podID="4407cffe-583e-4f8f-9da8-bc43ea8fcd32" containerID="bc56482707c3c180ac201c124976740fb1516b0515ecafde235c9e2e6ffa8cf9" exitCode=0 Oct 01 17:09:46 crc kubenswrapper[4726]: I1001 17:09:46.269323 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97p87" event={"ID":"4407cffe-583e-4f8f-9da8-bc43ea8fcd32","Type":"ContainerDied","Data":"bc56482707c3c180ac201c124976740fb1516b0515ecafde235c9e2e6ffa8cf9"} Oct 01 17:09:46 crc kubenswrapper[4726]: I1001 17:09:46.269540 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97p87" event={"ID":"4407cffe-583e-4f8f-9da8-bc43ea8fcd32","Type":"ContainerStarted","Data":"774e1b4a49c20083b8bc356400f670652ea7588e8728541c1bb2cd3f0d794c34"} Oct 01 17:09:48 crc kubenswrapper[4726]: I1001 17:09:48.290424 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97p87" event={"ID":"4407cffe-583e-4f8f-9da8-bc43ea8fcd32","Type":"ContainerStarted","Data":"e4428332308d2afb110026a723478ecc03423614a57028688d3cf25fb41fc83e"} Oct 01 17:09:49 crc kubenswrapper[4726]: I1001 17:09:49.299636 4726 generic.go:334] "Generic (PLEG): container finished" podID="4407cffe-583e-4f8f-9da8-bc43ea8fcd32" containerID="e4428332308d2afb110026a723478ecc03423614a57028688d3cf25fb41fc83e" exitCode=0 Oct 01 17:09:49 crc kubenswrapper[4726]: I1001 17:09:49.299958 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97p87" event={"ID":"4407cffe-583e-4f8f-9da8-bc43ea8fcd32","Type":"ContainerDied","Data":"e4428332308d2afb110026a723478ecc03423614a57028688d3cf25fb41fc83e"} Oct 01 17:09:50 crc kubenswrapper[4726]: I1001 17:09:50.312349 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97p87" event={"ID":"4407cffe-583e-4f8f-9da8-bc43ea8fcd32","Type":"ContainerStarted","Data":"9fe0e8e085b80fc796e121c79067e975d99629f7f17c8db1b7c2b7cd8094915c"} Oct 01 17:09:50 crc kubenswrapper[4726]: I1001 17:09:50.331458 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-97p87" podStartSLOduration=2.692693535 podStartE2EDuration="6.331442335s" podCreationTimestamp="2025-10-01 17:09:44 +0000 UTC" firstStartedPulling="2025-10-01 17:09:46.27067021 +0000 UTC m=+4239.172222787" lastFinishedPulling="2025-10-01 17:09:49.90941901 +0000 UTC m=+4242.810971587" observedRunningTime="2025-10-01 17:09:50.329294144 +0000 UTC m=+4243.230846721" watchObservedRunningTime="2025-10-01 17:09:50.331442335 +0000 UTC m=+4243.232994912" Oct 01 17:09:55 crc kubenswrapper[4726]: I1001 17:09:55.203951 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:55 crc kubenswrapper[4726]: I1001 17:09:55.204613 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:55 crc kubenswrapper[4726]: I1001 17:09:55.255751 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:55 crc kubenswrapper[4726]: I1001 17:09:55.439570 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:55 crc kubenswrapper[4726]: I1001 17:09:55.506394 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-97p87"] Oct 01 17:09:57 crc kubenswrapper[4726]: I1001 17:09:57.371805 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-97p87" podUID="4407cffe-583e-4f8f-9da8-bc43ea8fcd32" containerName="registry-server" containerID="cri-o://9fe0e8e085b80fc796e121c79067e975d99629f7f17c8db1b7c2b7cd8094915c" gracePeriod=2 Oct 01 17:09:57 crc kubenswrapper[4726]: I1001 17:09:57.944203 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.140413 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdfg5\" (UniqueName: \"kubernetes.io/projected/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-kube-api-access-sdfg5\") pod \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\" (UID: \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\") " Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.140674 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-utilities\") pod \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\" (UID: \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\") " Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.140692 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-catalog-content\") pod \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\" (UID: \"4407cffe-583e-4f8f-9da8-bc43ea8fcd32\") " Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.142583 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-utilities" (OuterVolumeSpecName: "utilities") pod "4407cffe-583e-4f8f-9da8-bc43ea8fcd32" (UID: "4407cffe-583e-4f8f-9da8-bc43ea8fcd32"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.150193 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-kube-api-access-sdfg5" (OuterVolumeSpecName: "kube-api-access-sdfg5") pod "4407cffe-583e-4f8f-9da8-bc43ea8fcd32" (UID: "4407cffe-583e-4f8f-9da8-bc43ea8fcd32"). InnerVolumeSpecName "kube-api-access-sdfg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.207881 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4407cffe-583e-4f8f-9da8-bc43ea8fcd32" (UID: "4407cffe-583e-4f8f-9da8-bc43ea8fcd32"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.242879 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.242913 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.242926 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdfg5\" (UniqueName: \"kubernetes.io/projected/4407cffe-583e-4f8f-9da8-bc43ea8fcd32-kube-api-access-sdfg5\") on node \"crc\" DevicePath \"\"" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.382119 4726 generic.go:334] "Generic (PLEG): container finished" podID="4407cffe-583e-4f8f-9da8-bc43ea8fcd32" containerID="9fe0e8e085b80fc796e121c79067e975d99629f7f17c8db1b7c2b7cd8094915c" exitCode=0 Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.382158 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97p87" event={"ID":"4407cffe-583e-4f8f-9da8-bc43ea8fcd32","Type":"ContainerDied","Data":"9fe0e8e085b80fc796e121c79067e975d99629f7f17c8db1b7c2b7cd8094915c"} Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.382176 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97p87" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.382195 4726 scope.go:117] "RemoveContainer" containerID="9fe0e8e085b80fc796e121c79067e975d99629f7f17c8db1b7c2b7cd8094915c" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.382183 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97p87" event={"ID":"4407cffe-583e-4f8f-9da8-bc43ea8fcd32","Type":"ContainerDied","Data":"774e1b4a49c20083b8bc356400f670652ea7588e8728541c1bb2cd3f0d794c34"} Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.418351 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-97p87"] Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.426591 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-97p87"] Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.431251 4726 scope.go:117] "RemoveContainer" containerID="e4428332308d2afb110026a723478ecc03423614a57028688d3cf25fb41fc83e" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.457225 4726 scope.go:117] "RemoveContainer" containerID="bc56482707c3c180ac201c124976740fb1516b0515ecafde235c9e2e6ffa8cf9" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.495350 4726 scope.go:117] "RemoveContainer" containerID="9fe0e8e085b80fc796e121c79067e975d99629f7f17c8db1b7c2b7cd8094915c" Oct 01 17:09:58 crc kubenswrapper[4726]: E1001 17:09:58.495808 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fe0e8e085b80fc796e121c79067e975d99629f7f17c8db1b7c2b7cd8094915c\": container with ID starting with 9fe0e8e085b80fc796e121c79067e975d99629f7f17c8db1b7c2b7cd8094915c not found: ID does not exist" containerID="9fe0e8e085b80fc796e121c79067e975d99629f7f17c8db1b7c2b7cd8094915c" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.495854 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe0e8e085b80fc796e121c79067e975d99629f7f17c8db1b7c2b7cd8094915c"} err="failed to get container status \"9fe0e8e085b80fc796e121c79067e975d99629f7f17c8db1b7c2b7cd8094915c\": rpc error: code = NotFound desc = could not find container \"9fe0e8e085b80fc796e121c79067e975d99629f7f17c8db1b7c2b7cd8094915c\": container with ID starting with 9fe0e8e085b80fc796e121c79067e975d99629f7f17c8db1b7c2b7cd8094915c not found: ID does not exist" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.495887 4726 scope.go:117] "RemoveContainer" containerID="e4428332308d2afb110026a723478ecc03423614a57028688d3cf25fb41fc83e" Oct 01 17:09:58 crc kubenswrapper[4726]: E1001 17:09:58.496231 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4428332308d2afb110026a723478ecc03423614a57028688d3cf25fb41fc83e\": container with ID starting with e4428332308d2afb110026a723478ecc03423614a57028688d3cf25fb41fc83e not found: ID does not exist" containerID="e4428332308d2afb110026a723478ecc03423614a57028688d3cf25fb41fc83e" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.496269 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4428332308d2afb110026a723478ecc03423614a57028688d3cf25fb41fc83e"} err="failed to get container status \"e4428332308d2afb110026a723478ecc03423614a57028688d3cf25fb41fc83e\": rpc error: code = NotFound desc = could not find container \"e4428332308d2afb110026a723478ecc03423614a57028688d3cf25fb41fc83e\": container with ID starting with e4428332308d2afb110026a723478ecc03423614a57028688d3cf25fb41fc83e not found: ID does not exist" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.496301 4726 scope.go:117] "RemoveContainer" containerID="bc56482707c3c180ac201c124976740fb1516b0515ecafde235c9e2e6ffa8cf9" Oct 01 17:09:58 crc kubenswrapper[4726]: E1001 17:09:58.496599 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc56482707c3c180ac201c124976740fb1516b0515ecafde235c9e2e6ffa8cf9\": container with ID starting with bc56482707c3c180ac201c124976740fb1516b0515ecafde235c9e2e6ffa8cf9 not found: ID does not exist" containerID="bc56482707c3c180ac201c124976740fb1516b0515ecafde235c9e2e6ffa8cf9" Oct 01 17:09:58 crc kubenswrapper[4726]: I1001 17:09:58.496625 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc56482707c3c180ac201c124976740fb1516b0515ecafde235c9e2e6ffa8cf9"} err="failed to get container status \"bc56482707c3c180ac201c124976740fb1516b0515ecafde235c9e2e6ffa8cf9\": rpc error: code = NotFound desc = could not find container \"bc56482707c3c180ac201c124976740fb1516b0515ecafde235c9e2e6ffa8cf9\": container with ID starting with bc56482707c3c180ac201c124976740fb1516b0515ecafde235c9e2e6ffa8cf9 not found: ID does not exist" Oct 01 17:09:59 crc kubenswrapper[4726]: I1001 17:09:59.819585 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4407cffe-583e-4f8f-9da8-bc43ea8fcd32" path="/var/lib/kubelet/pods/4407cffe-583e-4f8f-9da8-bc43ea8fcd32/volumes" Oct 01 17:10:23 crc kubenswrapper[4726]: I1001 17:10:23.413725 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 17:10:23 crc kubenswrapper[4726]: I1001 17:10:23.414193 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 17:10:53 crc kubenswrapper[4726]: I1001 17:10:53.413856 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 17:10:53 crc kubenswrapper[4726]: I1001 17:10:53.414519 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 17:11:23 crc kubenswrapper[4726]: I1001 17:11:23.413908 4726 patch_prober.go:28] interesting pod/machine-config-daemon-x5gdr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 17:11:23 crc kubenswrapper[4726]: I1001 17:11:23.414596 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 17:11:23 crc kubenswrapper[4726]: I1001 17:11:23.414655 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" Oct 01 17:11:23 crc kubenswrapper[4726]: I1001 17:11:23.415630 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f636fea95e8fc2020d3ea5c194d521a170b6d8d9b7ff2acf2fbaba2ddad73ed8"} pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 17:11:23 crc kubenswrapper[4726]: I1001 17:11:23.415747 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerName="machine-config-daemon" containerID="cri-o://f636fea95e8fc2020d3ea5c194d521a170b6d8d9b7ff2acf2fbaba2ddad73ed8" gracePeriod=600 Oct 01 17:11:23 crc kubenswrapper[4726]: E1001 17:11:23.546527 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:11:24 crc kubenswrapper[4726]: I1001 17:11:24.378337 4726 generic.go:334] "Generic (PLEG): container finished" podID="00d8cf08-6eff-420a-be98-e0dd45908be6" containerID="f636fea95e8fc2020d3ea5c194d521a170b6d8d9b7ff2acf2fbaba2ddad73ed8" exitCode=0 Oct 01 17:11:24 crc kubenswrapper[4726]: I1001 17:11:24.378454 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" event={"ID":"00d8cf08-6eff-420a-be98-e0dd45908be6","Type":"ContainerDied","Data":"f636fea95e8fc2020d3ea5c194d521a170b6d8d9b7ff2acf2fbaba2ddad73ed8"} Oct 01 17:11:24 crc kubenswrapper[4726]: I1001 17:11:24.379346 4726 scope.go:117] "RemoveContainer" containerID="6a39a92e4a63907159963d04757f77099dc1afa2ff471236819de46bcfbf3573" Oct 01 17:11:24 crc kubenswrapper[4726]: I1001 17:11:24.379824 4726 scope.go:117] "RemoveContainer" containerID="f636fea95e8fc2020d3ea5c194d521a170b6d8d9b7ff2acf2fbaba2ddad73ed8" Oct 01 17:11:24 crc kubenswrapper[4726]: E1001 17:11:24.380151 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:11:27 crc kubenswrapper[4726]: I1001 17:11:27.426736 4726 generic.go:334] "Generic (PLEG): container finished" podID="e49d5b8d-736e-440c-8cc1-66fd76db3937" containerID="2494b9dd9c22b6c332a37075f5b395a57d0d3a59674fbf19258d31b4e8122230" exitCode=0 Oct 01 17:11:27 crc kubenswrapper[4726]: I1001 17:11:27.426784 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lb9wc/must-gather-4pq84" event={"ID":"e49d5b8d-736e-440c-8cc1-66fd76db3937","Type":"ContainerDied","Data":"2494b9dd9c22b6c332a37075f5b395a57d0d3a59674fbf19258d31b4e8122230"} Oct 01 17:11:27 crc kubenswrapper[4726]: I1001 17:11:27.427442 4726 scope.go:117] "RemoveContainer" containerID="2494b9dd9c22b6c332a37075f5b395a57d0d3a59674fbf19258d31b4e8122230" Oct 01 17:11:28 crc kubenswrapper[4726]: I1001 17:11:28.411486 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lb9wc_must-gather-4pq84_e49d5b8d-736e-440c-8cc1-66fd76db3937/gather/0.log" Oct 01 17:11:36 crc kubenswrapper[4726]: I1001 17:11:36.809695 4726 scope.go:117] "RemoveContainer" containerID="f636fea95e8fc2020d3ea5c194d521a170b6d8d9b7ff2acf2fbaba2ddad73ed8" Oct 01 17:11:36 crc kubenswrapper[4726]: E1001 17:11:36.810348 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:11:39 crc kubenswrapper[4726]: I1001 17:11:39.901243 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lb9wc/must-gather-4pq84"] Oct 01 17:11:39 crc kubenswrapper[4726]: I1001 17:11:39.902212 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-lb9wc/must-gather-4pq84" podUID="e49d5b8d-736e-440c-8cc1-66fd76db3937" containerName="copy" containerID="cri-o://d03642d7648d13172802e303100b6660ba590556a25bc8172cc922264c903a4c" gracePeriod=2 Oct 01 17:11:39 crc kubenswrapper[4726]: I1001 17:11:39.913071 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lb9wc/must-gather-4pq84"] Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.363384 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lb9wc_must-gather-4pq84_e49d5b8d-736e-440c-8cc1-66fd76db3937/copy/0.log" Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.364434 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/must-gather-4pq84" Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.531653 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlgvr\" (UniqueName: \"kubernetes.io/projected/e49d5b8d-736e-440c-8cc1-66fd76db3937-kube-api-access-zlgvr\") pod \"e49d5b8d-736e-440c-8cc1-66fd76db3937\" (UID: \"e49d5b8d-736e-440c-8cc1-66fd76db3937\") " Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.531781 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e49d5b8d-736e-440c-8cc1-66fd76db3937-must-gather-output\") pod \"e49d5b8d-736e-440c-8cc1-66fd76db3937\" (UID: \"e49d5b8d-736e-440c-8cc1-66fd76db3937\") " Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.537542 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e49d5b8d-736e-440c-8cc1-66fd76db3937-kube-api-access-zlgvr" (OuterVolumeSpecName: "kube-api-access-zlgvr") pod "e49d5b8d-736e-440c-8cc1-66fd76db3937" (UID: "e49d5b8d-736e-440c-8cc1-66fd76db3937"). InnerVolumeSpecName "kube-api-access-zlgvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.558562 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lb9wc_must-gather-4pq84_e49d5b8d-736e-440c-8cc1-66fd76db3937/copy/0.log" Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.558982 4726 generic.go:334] "Generic (PLEG): container finished" podID="e49d5b8d-736e-440c-8cc1-66fd76db3937" containerID="d03642d7648d13172802e303100b6660ba590556a25bc8172cc922264c903a4c" exitCode=143 Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.559083 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lb9wc/must-gather-4pq84" Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.559046 4726 scope.go:117] "RemoveContainer" containerID="d03642d7648d13172802e303100b6660ba590556a25bc8172cc922264c903a4c" Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.621225 4726 scope.go:117] "RemoveContainer" containerID="2494b9dd9c22b6c332a37075f5b395a57d0d3a59674fbf19258d31b4e8122230" Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.636493 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlgvr\" (UniqueName: \"kubernetes.io/projected/e49d5b8d-736e-440c-8cc1-66fd76db3937-kube-api-access-zlgvr\") on node \"crc\" DevicePath \"\"" Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.732496 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e49d5b8d-736e-440c-8cc1-66fd76db3937-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e49d5b8d-736e-440c-8cc1-66fd76db3937" (UID: "e49d5b8d-736e-440c-8cc1-66fd76db3937"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.738252 4726 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e49d5b8d-736e-440c-8cc1-66fd76db3937-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.741039 4726 scope.go:117] "RemoveContainer" containerID="d03642d7648d13172802e303100b6660ba590556a25bc8172cc922264c903a4c" Oct 01 17:11:40 crc kubenswrapper[4726]: E1001 17:11:40.741580 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d03642d7648d13172802e303100b6660ba590556a25bc8172cc922264c903a4c\": container with ID starting with d03642d7648d13172802e303100b6660ba590556a25bc8172cc922264c903a4c not found: ID does not exist" containerID="d03642d7648d13172802e303100b6660ba590556a25bc8172cc922264c903a4c" Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.741667 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d03642d7648d13172802e303100b6660ba590556a25bc8172cc922264c903a4c"} err="failed to get container status \"d03642d7648d13172802e303100b6660ba590556a25bc8172cc922264c903a4c\": rpc error: code = NotFound desc = could not find container \"d03642d7648d13172802e303100b6660ba590556a25bc8172cc922264c903a4c\": container with ID starting with d03642d7648d13172802e303100b6660ba590556a25bc8172cc922264c903a4c not found: ID does not exist" Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.741750 4726 scope.go:117] "RemoveContainer" containerID="2494b9dd9c22b6c332a37075f5b395a57d0d3a59674fbf19258d31b4e8122230" Oct 01 17:11:40 crc kubenswrapper[4726]: E1001 17:11:40.742109 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2494b9dd9c22b6c332a37075f5b395a57d0d3a59674fbf19258d31b4e8122230\": container with ID starting with 2494b9dd9c22b6c332a37075f5b395a57d0d3a59674fbf19258d31b4e8122230 not found: ID does not exist" containerID="2494b9dd9c22b6c332a37075f5b395a57d0d3a59674fbf19258d31b4e8122230" Oct 01 17:11:40 crc kubenswrapper[4726]: I1001 17:11:40.742194 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2494b9dd9c22b6c332a37075f5b395a57d0d3a59674fbf19258d31b4e8122230"} err="failed to get container status \"2494b9dd9c22b6c332a37075f5b395a57d0d3a59674fbf19258d31b4e8122230\": rpc error: code = NotFound desc = could not find container \"2494b9dd9c22b6c332a37075f5b395a57d0d3a59674fbf19258d31b4e8122230\": container with ID starting with 2494b9dd9c22b6c332a37075f5b395a57d0d3a59674fbf19258d31b4e8122230 not found: ID does not exist" Oct 01 17:11:41 crc kubenswrapper[4726]: I1001 17:11:41.822392 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e49d5b8d-736e-440c-8cc1-66fd76db3937" path="/var/lib/kubelet/pods/e49d5b8d-736e-440c-8cc1-66fd76db3937/volumes" Oct 01 17:11:50 crc kubenswrapper[4726]: I1001 17:11:50.808674 4726 scope.go:117] "RemoveContainer" containerID="f636fea95e8fc2020d3ea5c194d521a170b6d8d9b7ff2acf2fbaba2ddad73ed8" Oct 01 17:11:50 crc kubenswrapper[4726]: E1001 17:11:50.809515 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:12:03 crc kubenswrapper[4726]: I1001 17:12:03.809435 4726 scope.go:117] "RemoveContainer" containerID="f636fea95e8fc2020d3ea5c194d521a170b6d8d9b7ff2acf2fbaba2ddad73ed8" Oct 01 17:12:03 crc kubenswrapper[4726]: E1001 17:12:03.810309 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:12:16 crc kubenswrapper[4726]: I1001 17:12:16.808885 4726 scope.go:117] "RemoveContainer" containerID="f636fea95e8fc2020d3ea5c194d521a170b6d8d9b7ff2acf2fbaba2ddad73ed8" Oct 01 17:12:16 crc kubenswrapper[4726]: E1001 17:12:16.811984 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" Oct 01 17:12:28 crc kubenswrapper[4726]: I1001 17:12:28.807600 4726 scope.go:117] "RemoveContainer" containerID="f636fea95e8fc2020d3ea5c194d521a170b6d8d9b7ff2acf2fbaba2ddad73ed8" Oct 01 17:12:28 crc kubenswrapper[4726]: E1001 17:12:28.808367 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x5gdr_openshift-machine-config-operator(00d8cf08-6eff-420a-be98-e0dd45908be6)\"" pod="openshift-machine-config-operator/machine-config-daemon-x5gdr" podUID="00d8cf08-6eff-420a-be98-e0dd45908be6" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067260410024446 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067260411017364 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067247357016525 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067247360015467 5ustar corecore